Science.gov

Sample records for audio feature space

  1. Digital audio authentication by robust feature embedding

    NASA Astrophysics Data System (ADS)

    Zmudzinski, Sascha; Munir, Badar; Steinebach, Martin

    2012-03-01

    We introduce an approach for verifying the integrity of digital audio recording by means of content-based integrity watermarking. Here an audio fingerprint is extracted from the Fourier domain and embedded as a digital watermark in the same domain. The design of the feature extraction allows a fine temporal resolution of the verification of the integrity. Experimental results show a good distinction between authentic and tampered audio content.

  2. Video salient event classification using audio features

    NASA Astrophysics Data System (ADS)

    Corchs, Silvia; Ciocca, Gianluigi; Fiori, Massimiliano; Gasparini, Francesca

    2014-03-01

    The aim of this work is to detect the events in video sequences that are salient with respect to the audio signal. In particular, we focus on the audio analysis of a video, with the goal of finding which are the significant features to detect audio-salient events. In our work we have extracted the audio tracks from videos of different sport events. For each video, we have manually labeled the salient audio-events using the binary markings. On each frame, features in both time and frequency domains have been considered. These features have been used to train different classifiers: Classification and Regression Trees, Support Vector Machine, and k-Nearest Neighbor. The classification performances are reported in terms of confusion matrices.

  3. Fall Detection Using Smartphone Audio Features.

    PubMed

    Cheffena, Michael

    2016-07-01

    An automated fall detection system based on smartphone audio features is developed. The spectrogram, mel frequency cepstral coefficents (MFCCs), linear predictive coding (LPC), and matching pursuit (MP) features of different fall and no-fall sound events are extracted from experimental data. Based on the extracted audio features, four different machine learning classifiers: k-nearest neighbor classifier (k-NN), support vector machine (SVM), least squares method (LSM), and artificial neural network (ANN) are investigated for distinguishing between fall and no-fall events. For each audio feature, the performance of each classifier in terms of sensitivity, specificity, accuracy, and computational complexity is evaluated. The best performance is achieved using spectrogram features with ANN classifier with sensitivity, specificity, and accuracy all above 98%. The classifier also has acceptable computational requirement for training and testing. The system is applicable in home environments where the phone is placed in the vicinity of the user. PMID:25915965

  4. Audio feature extraction using probability distribution function

    NASA Astrophysics Data System (ADS)

    Suhaib, A.; Wan, Khairunizam; Aziz, Azri A.; Hazry, D.; Razlan, Zuradzman M.; Shahriman A., B.

    2015-05-01

    Voice recognition has been one of the popular applications in robotic field. It is also known to be recently used for biometric and multimedia information retrieval system. This technology is attained from successive research on audio feature extraction analysis. Probability Distribution Function (PDF) is a statistical method which is usually used as one of the processes in complex feature extraction methods such as GMM and PCA. In this paper, a new method for audio feature extraction is proposed which is by using only PDF as a feature extraction method itself for speech analysis purpose. Certain pre-processing techniques are performed in prior to the proposed feature extraction method. Subsequently, the PDF result values for each frame of sampled voice signals obtained from certain numbers of individuals are plotted. From the experimental results obtained, it can be seen visually from the plotted data that each individuals' voice has comparable PDF values and shapes.

  5. Simple Solutions for Space Station Audio Problems

    NASA Technical Reports Server (NTRS)

    Wood, Eric

    2016-01-01

    Throughout this summer, a number of different projects were supported relating to various NASA programs, including the International Space Station (ISS) and Orion. The primary project that was worked on was designing and testing an acoustic diverter which could be used on the ISS to increase sound pressure levels in Node 1, a module that does not have any Audio Terminal Units (ATUs) inside it. This acoustic diverter is not intended to be a permanent solution to providing audio to Node 1; it is simply intended to improve conditions while more permanent solutions are under development. One of the most exciting aspects of this project is that the acoustic diverter is designed to be 3D printed on the ISS, using the 3D printer that was set up earlier this year. Because of this, no new hardware needs to be sent up to the station, and no extensive hardware testing needs to be performed on the ground before sending it to the station. Instead, the 3D part file can simply be uploaded to the station's 3D printer, where the diverter will be made.

  6. Audio-video feature correlation: faces and speech

    NASA Astrophysics Data System (ADS)

    Durand, Gwenael; Montacie, Claude; Caraty, Marie-Jose; Faudemay, Pascal

    1999-08-01

    This paper presents a study of the correlation of features automatically extracted from the audio stream and the video stream of audiovisual documents. In particular, we were interested in finding out whether speech analysis tools could be combined with face detection methods, and to what extend they should be combined. A generic audio signal partitioning algorithm as first used to detect Silence/Noise/Music/Speech segments in a full length movie. A generic object detection method was applied to the keyframes extracted from the movie in order to detect the presence or absence of faces. The correlation between the presence of a face in the keyframes and of the corresponding voice in the audio stream was studied. A third stream, which is the script of the movie, is warped on the speech channel in order to automatically label faces appearing in the keyframes with the name of the corresponding character. We naturally found that extracted audio and video features were related in many cases, and that significant benefits can be obtained from the joint use of audio and video analysis methods.

  7. Space Shuttle Orbiter audio subsystem. [to communication and tracking system

    NASA Technical Reports Server (NTRS)

    Stewart, C. H.

    1978-01-01

    The selection of the audio multiplex control configuration for the Space Shuttle Orbiter audio subsystem is discussed and special attention is given to the evaluation criteria of cost, weight and complexity. The specifications and design of the subsystem are described and detail is given to configurations of the audio terminal and audio central control unit (ATU, ACCU). The audio input from the ACCU, at a signal level of -12.2 to 14.8 dBV, nominal range, at 1 kHz, was found to have balanced source impedance and a balanced local impedance of 6000 + or - 600 ohms at 1 kHz, dc isolated. The Lyndon B. Johnson Space Center (JSC) electroacoustic test laboratory, an audio engineering facility consisting of a collection of acoustic test chambers, analyzed problems of speaker and headset performance, multiplexed control data coupled with audio channels, and the Orbiter cabin acoustic effects on the operational performance of voice communications. This system allows technical management and project engineering to address key constraining issues, such as identifying design deficiencies of the headset interface unit and the assessment of the Orbiter cabin performance of voice communications, which affect the subsystem development.

  8. Audio-visual synchrony and feature-selective attention co-amplify early visual processing.

    PubMed

    Keitel, Christian; Müller, Matthias M

    2016-05-01

    Our brain relies on neural mechanisms of selective attention and converging sensory processing to efficiently cope with rich and unceasing multisensory inputs. One prominent assumption holds that audio-visual synchrony can act as a strong attractor for spatial attention. Here, we tested for a similar effect of audio-visual synchrony on feature-selective attention. We presented two superimposed Gabor patches that differed in colour and orientation. On each trial, participants were cued to selectively attend to one of the two patches. Over time, spatial frequencies of both patches varied sinusoidally at distinct rates (3.14 and 3.63 Hz), giving rise to pulse-like percepts. A simultaneously presented pure tone carried a frequency modulation at the pulse rate of one of the two visual stimuli to introduce audio-visual synchrony. Pulsed stimulation elicited distinct time-locked oscillatory electrophysiological brain responses. These steady-state responses were quantified in the spectral domain to examine individual stimulus processing under conditions of synchronous versus asynchronous tone presentation and when respective stimuli were attended versus unattended. We found that both, attending to the colour of a stimulus and its synchrony with the tone, enhanced its processing. Moreover, both gain effects combined linearly for attended in-sync stimuli. Our results suggest that audio-visual synchrony can attract attention to specific stimulus features when stimuli overlap in space. PMID:26226930

  9. Electronystagmography and audio potentials in space flight.

    PubMed

    Thornton, W E; Biggers, W P; Thomas, W G; Pool, S L; Thagard, N E

    1985-08-01

    Beginning with the fourth flight of the Space Transport System (STS-4), objective measurements of inner ear function were conducted in near-zero G conditions in earth orbit. The problem of space motion sickness (SMS) was approached much like any disequilibrium problem encountered clinically. However, objective testing techniques had built-in limitations superimposed by the strict parameters inherent in each mission. An attempt was made to objectively characterize SMS, and to first ascertain whether the objective measurements indicated that this disorder was of peripheral or central origin. Electronystagmography and auditory brain stem response recordings were the primary investigative tools. One of the authors (W.E.T.) was a mission specialist on board the orbiter Challenger on the eight shuttle mission (STS-8) and had the opportunity to make direct and personal observations regarding SMS, an opportunity which has added immeasurably to our understanding of this disorder. Except for two abnormal ENG records, which remain to be explained, the remaining ENG records and all the ABR records made in the weightless environment of space were normal. PMID:4021685

  10. Electronystagmography and audio potentials in space flight

    NASA Technical Reports Server (NTRS)

    Thornton, William E.; Biggers, W. P.; Pool, Sam L.; Thomas, W. G.; Thagard, Norman E.

    1985-01-01

    Beginning with the fourth flight of the Space Transport System (STS-4), objective measurements of inner ear function were conducted in near-zero G conditions in earth orbit. The problem of space motion sickness (SMS) was approached much like any disequilibrium problem encountered clinically. However, objective testing techniques had built-in limitations superimposed by the strict parameters inherent in each mission. An attempt was made to objectively characterize SMS, and to first ascertain whether the objective measurements indicated that this disorder was of peripheral or central origin. Electronystagmography and auditory brain stem response recordings were the primary investigative tools. One of the authors (W.E.T.) was a mission specialist on board the orbiter Challenger on the eighth shuttle mission (STS-8) and had the opportunity to make direct and personal observations regarding SMS, an opportunity which has added immeasurably to our understanding of this disorder. Except for two abnormal ENG records, which remain to be explained, the remaining ENG records and all the ABR records made in the weightless environment of space were normal.

  11. Omnidirectional Audio-Visual Talker Localization Based on Dynamic Fusion of Audio-Visual Features Using Validity and Reliability Criteria

    NASA Astrophysics Data System (ADS)

    Denda, Yuki; Nishiura, Takanobu; Yamashita, Yoichi

    This paper proposes a robust omnidirectional audio-visual (AV) talker localizer for AV applications. The proposed localizer consists of two innovations. One of them is robust omnidirectional audio and visual features. The direction of arrival (DOA) estimation using an equilateral triangular microphone array, and human position estimation using an omnidirectional video camera extract the AV features. The other is a dynamic fusion of the AV features. The validity criterion, called the audioor visual-localization counter, validates each audio- or visual-feature. The reliability criterion, called the speech arriving evaluator, acts as a dynamic weight to eliminate any prior statistical properties from its fusion procedure. The proposed localizer can compatibly achieve talker localization in a speech activity and user localization in a non-speech activity under the identical fusion rule. Talker localization experiments were conducted in an actual room to evaluate the effectiveness of the proposed localizer. The results confirmed that the talker localization performance of the proposed AV localizer using the validity and reliability criteria is superior to that of conventional localizers.

  12. Environment Recognition for Digital Audio Forensics Using MPEG-7 and MEL Cepstral Features

    NASA Astrophysics Data System (ADS)

    Muhammad, Ghulam; Alghathbar, Khalid

    2011-07-01

    Environment recognition from digital audio for forensics application is a growing area of interest. However, compared to other branches of audio forensics, it is a less researched one. Especially less attention has been given to detect environment from files where foreground speech is present, which is a forensics scenario. In this paper, we perform several experiments focusing on the problems of environment recognition from audio particularly for forensics application. Experimental results show that the task is easier when audio files contain only environmental sound than when they contain both foreground speech and background environment. We propose a full set of MPEG-7 audio features combined with mel frequency cepstral coefficients (MFCCs) to improve the accuracy. In the experiments, the proposed approach significantly increases the recognition accuracy of environment sound even in the presence of high amount of foreground human speech.

  13. Audio-visual interaction and perceptual assessment of water features used over road traffic noise.

    PubMed

    Galbrun, Laurent; Calarco, Francesca M A

    2014-11-01

    This paper examines the audio-visual interaction and perception of water features used over road traffic noise, including their semantic aural properties, as well as their categorization and evocation properties. The research focused on a wide range of small to medium sized water features that can be used in gardens and parks to promote peacefulness and relaxation. Paired comparisons highlighted the inter-dependence between uni-modal (audio-only or visual-only) and bi-modal (audio-visual) perception, indicating that equal attention should be given to the design of both stimuli. In general, natural looking features tended to increase preference scores (compared to audio-only paired comparison scores), while manmade looking features decreased them. Semantic descriptors showed significant correlations with preferences and were found to be more reliable design criteria than physical parameters. A principal component analysis identified three components within the nine semantic attributes tested: "emotional assessment," "sound quality," and "envelopment and temporal variation." The first two showed significant correlations with audio-only preferences, "emotional assessment" being the most important predictor of preferences, and its attributes naturalness, relaxation, and freshness also being significantly correlated with preferences. Categorization results indicated that natural stream sounds are easily identifiable (unlike waterfalls and fountains), while evocation results showed no unique relationship with preferences. PMID:25373962

  14. Expansion Techniques of Embedding Audio Watermark Data Rate for Constructing Ubiquitous Acoustic Spaces

    NASA Astrophysics Data System (ADS)

    Modegi, Toshio

    We are proposing “Ubiquitous Acoustic Spaces”, where each sound source can emit some address information with audio signals and make us automatically access to its related cyber space, using handheld devices such as cellphones. In order to realize this concept, we have considered three types of extraction methods, which were an acoustic modulation, an audio fingerprint, and an audio watermark technique. Then we have proposed a novel audio watermarking technique, which enables contactless asynchronous detection of embedded audio watermarks through speaker and microphone devices. However its embedding data rate was around 10 [bps], which was not sufficient for embedding generally used URL address texts. Therefore, we have extended the embedding frequency range and proposed a duplicated embedding algorithm, which uses both previously proposed frequency division method and temporal division method together. By these improvements, possible embedding data rate could be extended to 61.5 [bps], and we could extract watermarks through public telephone networks, even from a cell phone sound source. In this paper, we describe abstracts of our improved watermark embedding and extracting algorithms, and experimental results of watermark extraction precision on several audio signal capturing conditions.

  15. Reconsidering the Role of Recorded Audio as a Rich, Flexible and Engaging Learning Space

    ERIC Educational Resources Information Center

    Middleton, Andrew

    2016-01-01

    Audio needs to be recognised as an integral medium capable of extending education's formal and informal, virtual and physical learning spaces. This paper reconsiders the value of educational podcasting through a review of literature and a module case study. It argues that a pedagogical understanding is needed and challenges technology-centred or…

  16. 47 CFR 25.214 - Technical requirements for space stations in the Satellite Digital Audio Radio Service and...

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... the Satellite Digital Audio Radio Service and associated terrestrial repeaters. 25.214 Section 25.214 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON CARRIER SERVICES SATELLITE COMMUNICATIONS Technical Standards § 25.214 Technical requirements for space stations in the Satellite Digital Audio...

  17. 47 CFR 25.214 - Technical requirements for space stations in the satellite digital audio radio service and...

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... the satellite digital audio radio service and associated terrestrial repeaters. 25.214 Section 25.214 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON CARRIER SERVICES SATELLITE COMMUNICATIONS Technical Standards § 25.214 Technical requirements for space stations in the satellite digital audio...

  18. 47 CFR 25.214 - Technical requirements for space stations in the satellite digital audio radio service and...

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... the satellite digital audio radio service and associated terrestrial repeaters. 25.214 Section 25.214 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON CARRIER SERVICES SATELLITE COMMUNICATIONS Technical Standards § 25.214 Technical requirements for space stations in the satellite digital audio...

  19. 47 CFR 25.214 - Technical requirements for space stations in the satellite digital audio radio service and...

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... the satellite digital audio radio service and associated terrestrial repeaters. 25.214 Section 25.214 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON CARRIER SERVICES SATELLITE COMMUNICATIONS Technical Standards § 25.214 Technical requirements for space stations in the satellite digital audio...

  20. 47 CFR 25.214 - Technical requirements for space stations in the Satellite Digital Audio Radio Service and...

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... the Satellite Digital Audio Radio Service and associated terrestrial repeaters. 25.214 Section 25.214 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON CARRIER SERVICES SATELLITE COMMUNICATIONS Technical Standards § 25.214 Technical requirements for space stations in the Satellite Digital Audio...

  1. Slow feature analysis with spiking neurons and its application to audio stimuli.

    PubMed

    Bellec, Guillaume; Galtier, Mathieu; Brette, Romain; Yger, Pierre

    2016-06-01

    Extracting invariant features in an unsupervised manner is crucial to perform complex computation such as object recognition, analyzing music or understanding speech. While various algorithms have been proposed to perform such a task, Slow Feature Analysis (SFA) uses time as a means of detecting those invariants, extracting the slowly time-varying components in the input signals. In this work, we address the question of how such an algorithm can be implemented by neurons, and apply it in the context of audio stimuli. We propose a projected gradient implementation of SFA that can be adapted to a Hebbian like learning rule dealing with biologically plausible neuron models. Furthermore, we show that a Spike-Timing Dependent Plasticity learning rule, shaped as a smoothed second derivative, implements SFA for spiking neurons. The theory is supported by numerical simulations, and to illustrate a simple use of SFA, we have applied it to auditory signals. We show that a single SFA neuron can learn to extract the tempo in sound recordings. PMID:27075919

  2. Transfer Learning across Feature-Rich Heterogeneous Feature Spaces via Feature-Space Remapping (FSR)

    PubMed Central

    Feuz, Kyle D.; Cook, Diane J.

    2016-01-01

    Transfer learning aims to improve performance on a target task by utilizing previous knowledge learned from source tasks. In this paper we introduce a novel heterogeneous transfer learning technique, Feature- Space Remapping (FSR), which transfers knowledge between domains with different feature spaces. This is accomplished without requiring typical feature-feature, feature instance, or instance-instance co-occurrence data. Instead we relate features in different feature-spaces through the construction of meta-features. We show how these techniques can utilize multiple source datasets to construct an ensemble learner which further improves performance. We apply FSR to an activity recognition problem and a document classification problem. The ensemble technique is able to outperform all other baselines and even performs better than a classifier trained using a large amount of labeled data in the target domain. These problems are especially difficult because in addition to having different feature-spaces, the marginal probability distributions and the class labels are also different. This work extends the state of the art in transfer learning by considering large transfer across dramatically different spaces. PMID:27019767

  3. Feature integration across space, time, and orientation

    PubMed Central

    Otto, Thomas U.; Öğmen, Haluk; Herzog, Michael H.

    2012-01-01

    The perception of a visual target can be strongly influenced by flanking stimuli. In static displays, performance on the target improves when the distance to the flanking elements increases- proposedly because feature pooling and integration vanishes with distance. Here, we studied feature integration with dynamic stimuli. We show that features of single elements presented within a continuous motion stream are integrated largely independent of spatial distance (and orientation). Hence, space based models of feature integration cannot be extended to dynamic stimuli. We suggest that feature integration is guided by perceptual grouping operations that maintain the identity of perceptual objects over space and time. PMID:19968428

  4. Feature space discriminant analysis for hyperspectral data feature reduction

    NASA Astrophysics Data System (ADS)

    Imani, Maryam; Ghassemian, Hassan

    2015-04-01

    Hyperspectral images contain a large number of spectral bands that allows us to distinguish different classes with more details. But, the number of available training samples is limited. Thus, feature reduction is an important step before classification of high dimensional data. Supervised feature extraction methods such as LDA, GDA, NWFE, and MMLDA use two criteria for feature reduction: between-class scatter and within-class scatter. We propose a supervised feature extraction method in this paper that uses a new criterion in addition to two mentioned measures. The proposed method, which is called feature space discriminant analysis (FSDA), at first, maximizes the between-spectral scatter matrix to increase the difference between extracted features. In the second step, FSDA, maximizes the between-class scatter matrix and minimizes the within-class scatter matrix simultaneously. The experimental results on five popular hyperspectral images show the better performance of FSDA in comparison with other supervised feature extraction methods in small sample size situation.

  5. Input space versus feature space in kernel-based methods.

    PubMed

    Schölkopf, B; Mika, S; Burges, C C; Knirsch, P; Müller, K R; Rätsch, G; Smola, A J

    1999-01-01

    This paper collects some ideas targeted at advancing our understanding of the feature spaces associated with support vector (SV) kernel functions. We first discuss the geometry of feature space. In particular, we review what is known about the shape of the image of input space under the feature space map, and how this influences the capacity of SV methods. Following this, we describe how the metric governing the intrinsic geometry of the mapped surface can be computed in terms of the kernel, using the example of the class of inhomogeneous polynomial kernels, which are often used in SV pattern recognition. We then discuss the connection between feature space and input space by dealing with the question of how one can, given some vector in feature space, find a preimage (exact or approximate) in input space. We describe algorithms to tackle this issue, and show their utility in two applications of kernel methods. First, we use it to reduce the computational complexity of SV decision functions; second, we combine it with the Kernel PCA algorithm, thereby constructing a nonlinear statistical denoising technique which is shown to perform well on real-world data. PMID:18252603

  6. Audio 2008: Audio Fixation

    ERIC Educational Resources Information Center

    Kaye, Alan L.

    2008-01-01

    Take a look around the bus or subway and see just how many people are bumping along to an iPod or an MP3 player. What they are listening to is their secret, but the many signature earbuds in sight should give one a real sense of just how pervasive digital audio has become. This article describes how that popularity is mirrored in library audio…

  7. Audio-visual imposture

    NASA Astrophysics Data System (ADS)

    Karam, Walid; Mokbel, Chafic; Greige, Hanna; Chollet, Gerard

    2006-05-01

    A GMM based audio visual speaker verification system is described and an Active Appearance Model with a linear speaker transformation system is used to evaluate the robustness of the verification. An Active Appearance Model (AAM) is used to automatically locate and track a speaker's face in a video recording. A Gaussian Mixture Model (GMM) based classifier (BECARS) is used for face verification. GMM training and testing is accomplished on DCT based extracted features of the detected faces. On the audio side, speech features are extracted and used for speaker verification with the GMM based classifier. Fusion of both audio and video modalities for audio visual speaker verification is compared with face verification and speaker verification systems. To improve the robustness of the multimodal biometric identity verification system, an audio visual imposture system is envisioned. It consists of an automatic voice transformation technique that an impostor may use to assume the identity of an authorized client. Features of the transformed voice are then combined with the corresponding appearance features and fed into the GMM based system BECARS for training. An attempt is made to increase the acceptance rate of the impostor and to analyzing the robustness of the verification system. Experiments are being conducted on the BANCA database, with a prospect of experimenting on the newly developed PDAtabase developed within the scope of the SecurePhone project.

  8. Feature Integration across Space, Time, and Orientation

    ERIC Educational Resources Information Center

    Otto, Thomas U.; Ogmen, Haluk; Herzog, Michael H.

    2009-01-01

    The perception of a visual target can be strongly influenced by flanking stimuli. In static displays, performance on the target improves when the distance to the flanking elements increases--presumably because feature pooling and integration vanishes with distance. Here, we studied feature integration with dynamic stimuli. We show that features of…

  9. Real World Audio

    NASA Technical Reports Server (NTRS)

    1998-01-01

    Crystal River Engineering was originally featured in Spinoff 1992 with the Convolvotron, a high speed digital audio processing system that delivers three-dimensional sound over headphones. The Convolvotron was developed for Ames' research on virtual acoustic displays. Crystal River is a now a subsidiary of Aureal Semiconductor, Inc. and they together develop and market the technology, which is a 3-D (three dimensional) audio technology known commercially today as Aureal 3D (A-3D). The technology has been incorporated into video games, surround sound systems, and sound cards.

  10. Efficient audio signal processing for embedded systems

    NASA Astrophysics Data System (ADS)

    Chiu, Leung Kin

    As mobile platforms continue to pack on more computational power, electronics manufacturers start to differentiate their products by enhancing the audio features. However, consumers also demand smaller devices that could operate for longer time, hence imposing design constraints. In this research, we investigate two design strategies that would allow us to efficiently process audio signals on embedded systems such as mobile phones and portable electronics. In the first strategy, we exploit properties of the human auditory system to process audio signals. We designed a sound enhancement algorithm to make piezoelectric loudspeakers sound ”richer" and "fuller." Piezoelectric speakers have a small form factor but exhibit poor response in the low-frequency region. In the algorithm, we combine psychoacoustic bass extension and dynamic range compression to improve the perceived bass coming out from the tiny speakers. We also developed an audio energy reduction algorithm for loudspeaker power management. The perceptually transparent algorithm extends the battery life of mobile devices and prevents thermal damage in speakers. This method is similar to audio compression algorithms, which encode audio signals in such a ways that the compression artifacts are not easily perceivable. Instead of reducing the storage space, however, we suppress the audio contents that are below the hearing threshold, therefore reducing the signal energy. In the second strategy, we use low-power analog circuits to process the signal before digitizing it. We designed an analog front-end for sound detection and implemented it on a field programmable analog array (FPAA). The system is an example of an analog-to-information converter. The sound classifier front-end can be used in a wide range of applications because programmable floating-gate transistors are employed to store classifier weights. Moreover, we incorporated a feature selection algorithm to simplify the analog front-end. A machine

  11. Oversampling the Minority Class in the Feature Space.

    PubMed

    Perez-Ortiz, Maria; Gutierrez, Pedro Antonio; Tino, Peter; Hervas-Martinez, Cesar

    2016-09-01

    The imbalanced nature of some real-world data is one of the current challenges for machine learning researchers. One common approach oversamples the minority class through convex combination of its patterns. We explore the general idea of synthetic oversampling in the feature space induced by a kernel function (as opposed to input space). If the kernel function matches the underlying problem, the classes will be linearly separable and synthetically generated patterns will lie on the minority class region. Since the feature space is not directly accessible, we use the empirical feature space (EFS) (a Euclidean space isomorphic to the feature space) for oversampling purposes. The proposed method is framed in the context of support vector machines, where the imbalanced data sets can pose a serious hindrance. The idea is investigated in three scenarios: 1) oversampling in the full and reduced-rank EFSs; 2) a kernel learning technique maximizing the data class separation to study the influence of the feature space structure (implicitly defined by the kernel function); and 3) a unified framework for preferential oversampling that spans some of the previous approaches in the literature. We support our investigation with extensive experiments over 50 imbalanced data sets. PMID:26316222

  12. pyAudioAnalysis: An Open-Source Python Library for Audio Signal Analysis

    PubMed Central

    Giannakopoulos, Theodoros

    2015-01-01

    Audio information plays a rather important role in the increasing digital content that is available today, resulting in a need for methodologies that automatically analyze such content: audio event recognition for home automations and surveillance systems, speech recognition, music information retrieval, multimodal analysis (e.g. audio-visual analysis of online videos for content-based recommendation), etc. This paper presents pyAudioAnalysis, an open-source Python library that provides a wide range of audio analysis procedures including: feature extraction, classification of audio signals, supervised and unsupervised segmentation and content visualization. pyAudioAnalysis is licensed under the Apache License and is available at GitHub (https://github.com/tyiannak/pyAudioAnalysis/). Here we present the theoretical background behind the wide range of the implemented methodologies, along with evaluation metrics for some of the methods. pyAudioAnalysis has been already used in several audio analysis research applications: smart-home functionalities through audio event detection, speech emotion recognition, depression classification based on audio-visual features, music segmentation, multimodal content-based movie recommendation and health applications (e.g. monitoring eating habits). The feedback provided from all these particular audio applications has led to practical enhancement of the library. PMID:26656189

  13. pyAudioAnalysis: An Open-Source Python Library for Audio Signal Analysis.

    PubMed

    Giannakopoulos, Theodoros

    2015-01-01

    Audio information plays a rather important role in the increasing digital content that is available today, resulting in a need for methodologies that automatically analyze such content: audio event recognition for home automations and surveillance systems, speech recognition, music information retrieval, multimodal analysis (e.g. audio-visual analysis of online videos for content-based recommendation), etc. This paper presents pyAudioAnalysis, an open-source Python library that provides a wide range of audio analysis procedures including: feature extraction, classification of audio signals, supervised and unsupervised segmentation and content visualization. pyAudioAnalysis is licensed under the Apache License and is available at GitHub (https://github.com/tyiannak/pyAudioAnalysis/). Here we present the theoretical background behind the wide range of the implemented methodologies, along with evaluation metrics for some of the methods. pyAudioAnalysis has been already used in several audio analysis research applications: smart-home functionalities through audio event detection, speech emotion recognition, depression classification based on audio-visual features, music segmentation, multimodal content-based movie recommendation and health applications (e.g. monitoring eating habits). The feedback provided from all these particular audio applications has led to practical enhancement of the library. PMID:26656189

  14. Multiclass Bayes error estimation by a feature space sampling technique

    NASA Technical Reports Server (NTRS)

    Mobasseri, B. G.; Mcgillem, C. D.

    1979-01-01

    A general Gaussian M-class N-feature classification problem is defined. An algorithm is developed that requires the class statistics as its only input and computes the minimum probability of error through use of a combined analytical and numerical integration over a sequence simplifying transformations of the feature space. The results are compared with those obtained by conventional techniques applied to a 2-class 4-feature discrimination problem with results previously reported and 4-class 4-feature multispectral scanner Landsat data classified by training and testing of the available data.

  15. Exploration of complex visual feature spaces for object perception

    PubMed Central

    Leeds, Daniel D.; Pyles, John A.; Tarr, Michael J.

    2014-01-01

    The mid- and high-level visual properties supporting object perception in the ventral visual pathway are poorly understood. In the absence of well-specified theory, many groups have adopted a data-driven approach in which they progressively interrogate neural units to establish each unit's selectivity. Such methods are challenging in that they require search through a wide space of feature models and stimuli using a limited number of samples. To more rapidly identify higher-level features underlying human cortical object perception, we implemented a novel functional magnetic resonance imaging method in which visual stimuli are selected in real-time based on BOLD responses to recently shown stimuli. This work was inspired by earlier primate physiology work, in which neural selectivity for mid-level features in IT was characterized using a simple parametric approach (Hung et al., 2012). To extend such work to human neuroimaging, we used natural and synthetic object stimuli embedded in feature spaces constructed on the basis of the complex visual properties of the objects themselves. During fMRI scanning, we employed a real-time search method to control continuous stimulus selection within each image space. This search was designed to maximize neural responses across a pre-determined 1 cm3 brain region within ventral cortex. To assess the value of this method for understanding object encoding, we examined both the behavior of the method itself and the complex visual properties the method identified as reliably activating selected brain regions. We observed: (1) Regions selective for both holistic and component object features and for a variety of surface properties; (2) Object stimulus pairs near one another in feature space that produce responses at the opposite extremes of the measured activity range. Together, these results suggest that real-time fMRI methods may yield more widely informative measures of selectivity within the broad classes of visual features

  16. The cross time and space features in remote sensing applications

    NASA Astrophysics Data System (ADS)

    Lu, J. X.; Song, W. L.; Qu, W.; Fu, J. E.; Pang, Z. G.

    2015-08-01

    Remote sensing is one subject of the modern geomatics, with a high priority for practical applications in which cross time and space analysis is one of its significant features. Object recognition and/or parameter retrieval are normally the first step in remote sensing applications, whereas cross time and space change analysis of those surface objects and/or parameters will make remote sensing applications more valuable. Based on a short review on the historic evolution of remote sensing and its current classification system, the cross time and space features commonly existing in remote sensing applications were discussed. The paper, aiming at improving remote sensing applications and promoting development of the remote sensing subject from a new vision, proposed a methodology based subject classification approach for remote sensing and then suggest to establish the theory of cross time and space remote sensing applications. The authors believe that such a new cross time and space concept meets the demand for new theories and new ideas from remote sensing subject and is of practical help to future remote sensing applications.

  17. Features of the Gravity Probe B Space Vehicle

    NASA Astrophysics Data System (ADS)

    Reeve, William; Green, Gaylord

    2007-04-01

    Space vehicle performance enabled successful relativity data collection throughout the Gravity Probe B mission. Precision pointing and drag-free translation control was maintained using proportional helium micro-thrusters. Electrical power was provided by rigid, double sided solar arrays. The 1.8 kelvin science instrument temperature was maintained using the largest cryogenic liquid helium dewar ever flown in space. The flight software successfully performed autonomous operations and safemode protection. Features of the Gravity Probe B Space Vehicle mechanisms include: 1) sixteen helium micro-thrusters, the first proportional thrusters flown in space, and large-orifice thruster isolation valves, 2) seven precision and high-authority mass trim mechanisms, 3) four non-pyrotechnic, highly reliable solar array deployment and release mechanism sets. Early incremental prototyping was used extensively to reduce spacecraft development risk. All spacecraft systems were redundant and provided multiple failure tolerance in critical systems. Lockheed Martin performed the spacecraft design, systems engineering, hardware and software integration, environmental testing and launch base operations, as well as on-orbit operations support for the Gravity Probe B space science experiment.

  18. Feature extraction on local jet space for texture classification

    NASA Astrophysics Data System (ADS)

    Oliveira, Marcos William da Silva; da Silva, Núbia Rosa; Manzanera, Antoine; Bruno, Odemir Martinez

    2015-12-01

    The proposal of this study is to analyze the texture pattern recognition over the local jet space looking forward to improve the texture characterization. Local jets decompose the image based on partial derivatives allowing the texture feature extraction be exploited in different levels of geometrical structures. Each local jet component evidences a different local pattern, such as, flat regions, directional variations and concavity or convexity. Subsequently, a texture descriptor is used to extract features from 0th, 1st and 2nd-derivative components. Four well-known databases (Brodatz, Vistex, Usptex and Outex) and four texture descriptors (Fourier descriptors, Gabor filters, Local Binary Pattern and Local Binary Pattern Variance) were used to validate the idea, showing in most cases an increase of the success rates.

  19. Detecting double compression of audio signal

    NASA Astrophysics Data System (ADS)

    Yang, Rui; Shi, Yun Q.; Huang, Jiwu

    2010-01-01

    MP3 is the most popular audio format nowadays in our daily life, for example music downloaded from the Internet and file saved in the digital recorder are often in MP3 format. However, low bitrate MP3s are often transcoded to high bitrate since high bitrate ones are of high commercial value. Also audio recording in digital recorder can be doctored easily by pervasive audio editing software. This paper presents two methods for the detection of double MP3 compression. The methods are essential for finding out fake-quality MP3 and audio forensics. The proposed methods use support vector machine classifiers with feature vectors formed by the distributions of the first digits of the quantized MDCT (modified discrete cosine transform) coefficients. Extensive experiments demonstrate the effectiveness of the proposed methods. To the best of our knowledge, this piece of work is the first one to detect double compression of audio signal.

  20. Feature extraction algorithm for space targets based on fractal theory

    NASA Astrophysics Data System (ADS)

    Tian, Balin; Yuan, Jianping; Yue, Xiaokui; Ning, Xin

    2007-11-01

    In order to offer a potential for extending the life of satellites and reducing the launch and operating costs, satellite servicing including conducting repairs, upgrading and refueling spacecraft on-orbit become much more frequently. Future space operations can be more economically and reliably executed using machine vision systems, which can meet real time and tracking reliability requirements for image tracking of space surveillance system. Machine vision was applied to the research of relative pose for spacecrafts, the feature extraction algorithm was the basis of relative pose. In this paper fractal geometry based edge extraction algorithm which can be used in determining and tracking the relative pose of an observed satellite during proximity operations in machine vision system was presented. The method gets the gray-level image distributed by fractal dimension used the Differential Box-Counting (DBC) approach of the fractal theory to restrain the noise. After this, we detect the consecutive edge using Mathematical Morphology. The validity of the proposed method is examined by processing and analyzing images of space targets. The edge extraction method not only extracts the outline of the target, but also keeps the inner details. Meanwhile, edge extraction is only processed in moving area to reduce computation greatly. Simulation results compared edge detection using the method which presented by us with other detection methods. The results indicate that the presented algorithm is a valid method to solve the problems of relative pose for spacecrafts.

  1. Enhancing Manual Scan Registration Using Audio Cues

    NASA Astrophysics Data System (ADS)

    Ntsoko, T.; Sithole, G.

    2014-04-01

    Indoor mapping and modelling requires that acquired data be processed by editing, fusing, formatting the data, amongst other operations. Currently the manual interaction the user has with the point cloud (data) while processing it is visual. Visual interaction does have limitations, however. One way of dealing with these limitations is to augment audio in point cloud processing. Audio augmentation entails associating points of interest in the point cloud with audio objects. In coarse scan registration, reverberation, intensity and frequency audio cues were exploited to help the user estimate depth and occupancy of space of points of interest. Depth estimations were made reliably well when intensity and frequency were both used as depth cues. Coarse changes of depth could be estimated in this manner. The depth between surfaces can therefore be estimated with the aid of the audio objects. Sound reflections of an audio object provided reliable information of the object surroundings in some instances. For a point/area of interest in the point cloud, these reflections can be used to determine the unseen events around that point/area of interest. Other processing techniques could benefit from this while other information is estimated using other audio cues like binaural cues and Head Related Transfer Functions. These other cues could be used in position estimations of audio objects to aid in problems such as indoor navigation problems.

  2. Web Audio/Video Streaming Tool

    NASA Technical Reports Server (NTRS)

    Guruvadoo, Eranna K.

    2003-01-01

    In order to promote NASA-wide educational outreach program to educate and inform the public of space exploration, NASA, at Kennedy Space Center, is seeking efficient ways to add more contents to the web by streaming audio/video files. This project proposes a high level overview of a framework for the creation, management, and scheduling of audio/video assets over the web. To support short-term goals, the prototype of a web-based tool is designed and demonstrated to automate the process of streaming audio/video files. The tool provides web-enabled users interfaces to manage video assets, create publishable schedules of video assets for streaming, and schedule the streaming events. These operations are performed on user-defined and system-derived metadata of audio/video assets stored in a relational database while the assets reside on separate repository. The prototype tool is designed using ColdFusion 5.0.

  3. Space.

    ERIC Educational Resources Information Center

    Web Feet K-8, 2001

    2001-01-01

    This annotated subject guide to Web sites and additional resources focuses on space and astronomy. Specifies age levels for resources that include Web sites, CD-ROMS and software, videos, books, audios, and magazines; offers professional resources; and presents a relevant class activity. (LRW)

  4. Trajectory analysis via a geometric feature space approach

    SciTech Connect

    Rintoul, Mark D.; Wilson, Andrew T.

    2015-10-05

    This study aimed to organize a body of trajectories in order to identify, search for and classify both common and uncommon behaviors among objects such as aircraft and ships. Existing comparison functions such as the Fréchet distance are computationally expensive and yield counterintuitive results in some cases. We propose an approach using feature vectors whose components represent succinctly the salient information in trajectories. These features incorporate basic information such as the total distance traveled and the distance between start/stop points as well as geometric features related to the properties of the convex hull, trajectory curvature and general distance geometry. Additionally, these features can generally be mapped easily to behaviors of interest to humans who are searching large databases. Most of these geometric features are invariant under rigid transformation. Furthermore, we demonstrate the use of different subsets of these features to identify trajectories similar to an exemplar, cluster a database of several hundred thousand trajectories and identify outliers.

  5. MedlinePlus FAQ: Is audio description available for videos on MedlinePlus?

    MedlinePlus

    ... audiodescription.html Question: Is audio description available for videos on MedlinePlus? To use the sharing features on ... page, please enable JavaScript. Answer: Audio description of videos helps make the content of videos accessible to ...

  6. Aeronautical audio broadcasting via satellite

    NASA Astrophysics Data System (ADS)

    Tzeng, Forrest F.

    A system design for aeronautical audio broadcasting, with C-band uplink and L-band downlink, via Inmarsat space segments is presented. Near-transparent-quality compression of 5-kHz bandwidth audio at 20.5 kbit/s is achieved based on a hybrid technique employing linear predictive modeling and transform-domain residual quantization. Concatenated Reed-Solomon/convolutional codes with quadrature phase shift keying are selected for bandwidth and power efficiency. RF bandwidth at 25 kHz per channel, and a decoded bit error rate at 10(exp -6) with E(sub b)/N(sub o) at 3.75 dB are obtained. An interleaver, scrambler, modem synchronization, and frame format were designed, and frequency-division multiple access was selected over code-division multiple access. A link budget computation based on a worst-case scenario indicates sufficient system power margins. Transponder occupancy analysis for 72 audio channels demonstrates ample remaining capacity to accommodate emerging aeronautical services.

  7. Aeronautical audio broadcasting via satellite

    NASA Technical Reports Server (NTRS)

    Tzeng, Forrest F.

    1993-01-01

    A system design for aeronautical audio broadcasting, with C-band uplink and L-band downlink, via Inmarsat space segments is presented. Near-transparent-quality compression of 5-kHz bandwidth audio at 20.5 kbit/s is achieved based on a hybrid technique employing linear predictive modeling and transform-domain residual quantization. Concatenated Reed-Solomon/convolutional codes with quadrature phase shift keying are selected for bandwidth and power efficiency. RF bandwidth at 25 kHz per channel, and a decoded bit error rate at 10(exp -6) with E(sub b)/N(sub o) at 3.75 dB are obtained. An interleaver, scrambler, modem synchronization, and frame format were designed, and frequency-division multiple access was selected over code-division multiple access. A link budget computation based on a worst-case scenario indicates sufficient system power margins. Transponder occupancy analysis for 72 audio channels demonstrates ample remaining capacity to accommodate emerging aeronautical services.

  8. Forensic audio watermark detection

    NASA Astrophysics Data System (ADS)

    Steinebach, Martin; Zmudzinski, Sascha; Petrautzki, Dirk

    2012-03-01

    Digital audio watermarking detection is often computational complex and requires at least as much audio information as required to embed a complete watermark. In some applications, especially real-time monitoring, this is an important drawback. The reason for this is the usage of sync sequences at the beginning of the watermark, allowing a decision about the presence only if at least the sync has been found and retrieved. We propose an alternative method for detecting the presence of a watermark. Based on the knowledge of the secret key used for embedding, we create a mark for all potential marking stages and then use a sliding window to test a given audio file on the presence of statistical characteristics caused by embedding. In this way we can detect a watermark in less than 1 second of audio.

  9. Audio signal processor

    NASA Technical Reports Server (NTRS)

    Hymer, R. L.

    1970-01-01

    System provides automatic volume control for an audio amplifier or a voice communication system without introducing noise surges during pauses in the input, and without losing the initial signal when the input resumes.

  10. Detecting Image Splicing Using Merged Features in Chroma Space

    PubMed Central

    Liu, Guangjie; Dai, Yuewei

    2014-01-01

    Image splicing is an image editing method to copy a part of an image and paste it onto another image, and it is commonly followed by postprocessing such as local/global blurring, compression, and resizing. To detect this kind of forgery, the image rich models, a feature set successfully used in the steganalysis is evaluated on the splicing image dataset at first, and the dominant submodel is selected as the first kind of feature. The selected feature and the DCT Markov features are used together to detect splicing forgery in the chroma channel, which is convinced effective in splicing detection. The experimental results indicate that the proposed method can detect splicing forgeries with lower error rate compared to the previous literature. PMID:24574877

  11. Trajectory analysis via a geometric feature space approach

    DOE PAGESBeta

    Rintoul, Mark D.; Wilson, Andrew T.

    2015-10-05

    This study aimed to organize a body of trajectories in order to identify, search for and classify both common and uncommon behaviors among objects such as aircraft and ships. Existing comparison functions such as the Fréchet distance are computationally expensive and yield counterintuitive results in some cases. We propose an approach using feature vectors whose components represent succinctly the salient information in trajectories. These features incorporate basic information such as the total distance traveled and the distance between start/stop points as well as geometric features related to the properties of the convex hull, trajectory curvature and general distance geometry. Additionally,more » these features can generally be mapped easily to behaviors of interest to humans who are searching large databases. Most of these geometric features are invariant under rigid transformation. Furthermore, we demonstrate the use of different subsets of these features to identify trajectories similar to an exemplar, cluster a database of several hundred thousand trajectories and identify outliers.« less

  12. Limitations in 4-Year-Old Children's Sensitivity to the Spacing among Facial Features

    ERIC Educational Resources Information Center

    Mondloch, Catherine J.; Thomson, Kendra

    2008-01-01

    Four-year-olds' sensitivity to differences among faces in the spacing of features was tested under 4 task conditions: judging distinctiveness when the external contour was visible and when it was occluded, simultaneous match-to-sample, and recognizing the face of a friend. In each task, the foil differed only in the spacing of features, and…

  13. Space-based RF signal classification using adaptive wavelet features

    SciTech Connect

    Caffrey, M.; Briles, S.

    1995-04-01

    RF signals are dispersed in frequency as they propagate through the ionosphere. For wide-band signals, this results in nonlinearly- chirped-frequency, transient signals in the VHF portion of the spectrum. This ionospheric dispersion provide a means of discriminating wide-band transients from other signals (e.g., continuous-wave carriers, burst communications, chirped-radar signals, etc.). The transient nature of these dispersed signals makes them candidates for wavelet feature selection. Rather than choosing a wavelet ad hoc, we adaptively compute an optimal mother wavelet via a neural network. Gaussian weighted, linear frequency modulate (GLFM) wavelets are linearly combined by the network to generate our application specific mother wavelet, which is optimized for its capacity to select features that discriminate between the dispersed signals and clutter (e.g., multiple continuous-wave carriers), not for its ability to represent the dispersed signal. The resulting mother wavelet is then used to extract features for a neutral network classifier. The performance of the adaptive wavelet classifier is the compared to an FFT based neural network classifier.

  14. Agency Video, Audio and Imagery Library

    NASA Technical Reports Server (NTRS)

    Grubbs, Rodney

    2015-01-01

    The purpose of this presentation was to inform the ISS International Partners of the new NASA Agency Video, Audio and Imagery Library (AVAIL) website. AVAIL is a new resource for the public to search for and download NASA-related imagery, and is not intended to replace the current process by which the International Partners receive their Space Station imagery products.

  15. Reverse engineering the face space: Discovering the critical features for face identification.

    PubMed

    Abudarham, Naphtali; Yovel, Galit

    2016-01-01

    How do we identify people? What are the critical facial features that define an identity and determine whether two faces belong to the same person or different people? To answer these questions, we applied the face space framework, according to which faces are represented as points in a multidimensional feature space, such that face space distances are correlated with perceptual similarities between faces. In particular, we developed a novel method that allowed us to reveal the critical dimensions (i.e., critical features) of the face space. To that end, we constructed a concrete face space, which included 20 facial features of natural face images, and asked human observers to evaluate feature values (e.g., how thick are the lips). Next, we systematically and quantitatively changed facial features, and measured the perceptual effects of these manipulations. We found that critical features were those for which participants have high perceptual sensitivity (PS) for detecting differences across identities (e.g., which of two faces has thicker lips). Furthermore, these high PS features vary minimally across different views of the same identity, suggesting high PS features support face recognition across different images of the same face. The methods described here set an infrastructure for discovering the critical features of other face categories not studied here (e.g., Asians, familiar) as well as other aspects of face processing, such as attractiveness or trait inferences. PMID:26928056

  16. The method of narrow-band audio classification based on universal noise background model

    NASA Astrophysics Data System (ADS)

    Rui, Rui; Bao, Chang-chun

    2013-03-01

    Audio classification is the basis of content-based audio analysis and retrieval. The conventional classification methods mainly depend on feature extraction of audio clip, which certainly increase the time requirement for classification. An approach for classifying the narrow-band audio stream based on feature extraction of audio frame-level is presented in this paper. The audio signals are divided into speech, instrumental music, song with accompaniment and noise using the Gaussian mixture model (GMM). In order to satisfy the demand of actual environment changing, a universal noise background model (UNBM) for white noise, street noise, factory noise and car interior noise is built. In addition, three feature schemes are considered to optimize feature selection. The experimental results show that the proposed algorithm achieves a high accuracy for audio classification, especially under each noise background we used and keep the classification time less than one second.

  17. Enhancing Navigation Skills through Audio Gaming

    PubMed Central

    Sánchez, Jaime; Sáenz, Mauricio; Pascual-Leone, Alvaro; Merabet, Lotfi

    2014-01-01

    We present the design, development and initial cognitive evaluation of an Audio-based Environment Simulator (AbES). This software allows a blind user to navigate through a virtual representation of a real space for the purposes of training orientation and mobility skills. Our findings indicate that users feel satisfied and self-confident when interacting with the audio-based interface, and the embedded sounds allow them to correctly orient themselves and navigate within the virtual world. Furthermore, users are able to transfer spatial information acquired through virtual interactions into real world navigation and problem solving tasks. PMID:25505796

  18. 3D Audio System

    NASA Technical Reports Server (NTRS)

    1992-01-01

    Ames Research Center research into virtual reality led to the development of the Convolvotron, a high speed digital audio processing system that delivers three-dimensional sound over headphones. It consists of a two-card set designed for use with a personal computer. The Convolvotron's primary application is presentation of 3D audio signals over headphones. Four independent sound sources are filtered with large time-varying filters that compensate for motion. The perceived location of the sound remains constant. Possible applications are in air traffic control towers or airplane cockpits, hearing and perception research and virtual reality development.

  19. Digital Multicasting of Multiple Audio Streams

    NASA Technical Reports Server (NTRS)

    Macha, Mitchell; Bullock, John

    2007-01-01

    at the MCC. In the other access-control provision, the program verifies that the user is authorized to have access to the audio streams. Once both access-control checks are completed, the audio software presents a graphical display that includes audiostream-selection buttons and volume-control sliders. The user can select all or any subset of the available audio streams and can adjust the volume of each stream independently of that of the other streams. The audio-player program spawns a "read" process for the selected stream(s). The spawned process sends, to the router(s), a "multicast-join" request for the selected streams. The router(s) responds to the request by sending the encrypted multicast packets to the spawned process. The spawned process receives the encrypted multicast packets and sends a decryption packet to audio-driver software. As the volume or muting features are changed by the user, interrupts are sent to the spawned process to change the corresponding attributes sent to the audio-driver software. The total latency of this system - that is, the total time from the origination of the audio signals to generation of sound at a listener s computer - lies between four and six seconds.

  20. Towards Semantically Sensitive Text Clustering: A Feature Space Modeling Technology Based on Dimension Extension

    PubMed Central

    Liu, Yuanchao; Liu, Ming; Wang, Xin

    2015-01-01

    The objective of text clustering is to divide document collections into clusters based on the similarity between documents. In this paper, an extension-based feature modeling approach towards semantically sensitive text clustering is proposed along with the corresponding feature space construction and similarity computation method. By combining the similarity in traditional feature space and that in extension space, the adverse effects of the complexity and diversity of natural language can be addressed and clustering semantic sensitivity can be improved correspondingly. The generated clusters can be organized using different granularities. The experimental evaluations on well-known clustering algorithms and datasets have verified the effectiveness of our approach. PMID:25794172

  1. Audio Feedback -- Better Feedback?

    ERIC Educational Resources Information Center

    Voelkel, Susanne; Mello, Luciane V.

    2014-01-01

    National Student Survey (NSS) results show that many students are dissatisfied with the amount and quality of feedback they get for their work. This study reports on two case studies in which we tried to address these issues by introducing audio feedback to one undergraduate (UG) and one postgraduate (PG) class, respectively. In case study one…

  2. Wavelet-based audio embedding and audio/video compression

    NASA Astrophysics Data System (ADS)

    Mendenhall, Michael J.; Claypoole, Roger L., Jr.

    2001-12-01

    Watermarking, traditionally used for copyright protection, is used in a new and exciting way. An efficient wavelet-based watermarking technique embeds audio information into a video signal. Several effective compression techniques are applied to compress the resulting audio/video signal in an embedded fashion. This wavelet-based compression algorithm incorporates bit-plane coding, index coding, and Huffman coding. To demonstrate the potential of this audio embedding and audio/video compression algorithm, we embed an audio signal into a video signal and then compress. Results show that overall compression rates of 15:1 can be achieved. The video signal is reconstructed with a median PSNR of nearly 33 dB. Finally, the audio signal is extracted from the compressed audio/video signal without error.

  3. Systematic acquisition of audio classes for elevator surveillance

    NASA Astrophysics Data System (ADS)

    Radhakrishnan, Regunathan; Divakaran, Ajay

    2005-03-01

    We present a systematic framework for arriving at audio classes for detection of crimes in elevators. We use a time series analysis framework to analyze the low-level features extracted from the audio of an elevator surveillance content to perform an inlier/outlier based temporal segmentation. Since suspicious events in elevators are outliers in a background of usual events, such a segmentation help bring out such events without any a priori knowledge. Then, by performing an automatic clustering on the detected outliers, we identify consistent patterns for which we can train supervised detectors. We apply the proposed framework to a collection of elevator surveillance audio data to systematically acquire audio classes such as banging, footsteps, non-neutral speech and normal speech etc. Based on the observation that the banging audio class and non-neutral speech class are indicative of suspicious events in the elevator data set, we are able to detect all of the suspicious activities without any misses.

  4. 3D Point Correspondence by Minimum Description Length in Feature Space.

    PubMed

    Chen, Jiun-Hung; Zheng, Ke Colin; Shapiro, Linda G

    2010-01-01

    Finding point correspondences plays an important role in automatically building statistical shape models from a training set of 3D surfaces. For the point correspondence problem, Davies et al. [1] proposed a minimum-description-length-based objective function to balance the training errors and generalization ability. A recent evaluation study [2] that compares several well-known 3D point correspondence methods for modeling purposes shows that the MDL-based approach [1] is the best method. We adapt the MDL-based objective function for a feature space that can exploit nonlinear properties in point correspondences, and propose an efficient optimization method to minimize the objective function directly in the feature space, given that the inner product of any vector pair can be computed in the feature space. We further employ a Mercer kernel [3] to define the feature space implicitly. A key aspect of our proposed framework is the generalization of the MDL-based objective function to kernel principal component analysis (KPCA) [4] spaces and the design of a gradient-descent approach to minimize such an objective function. We compare the generalized MDL objective function on KPCA spaces with the original one and evaluate their abilities in terms of reconstruction errors and specificity. From our experimental results on different sets of 3D shapes of human body organs, the proposed method performs significantly better than the original method. PMID:25328917

  5. A scale space feature based registration technique for fusion of satellite imagery

    NASA Technical Reports Server (NTRS)

    Raghavan, Srini; Cromp, Robert F.; Campbell, William C.

    1997-01-01

    Feature based registration is one of the most reliable methods to register multi-sensor images (both active and passive imagery) since features are often more reliable than intensity or radiometric values. The only situation where a feature based approach will fail is when the scene is completely homogenous or densely textural in which case a combination of feature and intensity based methods may yield better results. In this paper, we present some preliminary results of testing our scale space feature based registration technique, a modified version of feature based method developed earlier for classification of multi-sensor imagery. The proposed approach removes the sensitivity in parameter selection experienced in the earlier version as explained later.

  6. Audio distribution and Monitoring Circuit

    NASA Technical Reports Server (NTRS)

    Kirkland, J. M.

    1983-01-01

    Versatile circuit accepts and distributes TV audio signals. Three-meter audio distribution and monitoring circuit provides flexibility in monitoring, mixing, and distributing audio inputs and outputs at various signal and impedance levels. Program material is simultaneously monitored on three channels, or single-channel version built to monitor transmitted or received signal levels, drive speakers, interface to building communications, and drive long-line circuits.

  7. Hiding Data in Audio Signal

    NASA Astrophysics Data System (ADS)

    Bhattacharyya, Debnath; Dutta, Poulami; Balitanas, Maricel O.; Kim, Tai-Hoon; Das, Purnendu

    This paper describes the LSB technique for secure data transfer. Secret information can be hidden inside all sorts of cover information: text, images, audio, video and more. Embedding secret messages in digital sound is usually a more difficult process. Varieties of techniques for embedding information in digital audio have been established. These are parity coding, phase coding, spread spectrum, echo hiding, LSB. Least significant bits (LSB) insertion is one of the simplest approaches to embedding information in audio file.

  8. Quantitative lithologic mapping in spectral ratio feature space - Volcanic, sedimentary and metamorphic terrains

    NASA Technical Reports Server (NTRS)

    Campos-Marquetti, Raul, Jr.; Rockwell, Barnaby

    1990-01-01

    The nature of spectral lithologic mapping is studied utilizing ratios centered around the wavelength means of TM imagery. Laboratory-derived spectra are analyzed to determine the two-dimensional relationships and distributions visible in spectral ratio feature space. The spectral distributions of various rocks and minerals in ratio feature space are found to be controlled by several spectrally dominant molecules. Three study areas were examined: Rawhide Mining District, Nevada; Manzano Mountains, New Mexico; and the Sevilleta Long Term Ecological Research site in New Mexico. It is shown that, in the comparison of two ratio plots of laboratory reflectance spectra, i.e., 0.66/0.485 micron versus 1.65/2.22 microns with those derived from TM data, several molecules spectrally dominate the reflectance characteristic of surface lithologic units. Utilizing the above ratio combination, two areas are successfully mapped based on their distribution in spectral ratio feature space.

  9. Teledesic Global Wireless Broadband Network: Space Infrastructure Architecture, Design Features and Technologies

    NASA Technical Reports Server (NTRS)

    Stuart, James R.

    1995-01-01

    The Teledesic satellites are a new class of small satellites which demonstrate the important commercial benefits of using technologies developed for other purposes by U.S. National Laboratories. The Teledesic satellite architecture, subsystem design features, and new technologies are described. The new Teledesic satellite manufacturing, integration, and test approaches which use modern high volume production techniques and result in surprisingly low space segment costs are discussed. The constellation control and management features and attendant software architecture features are addressed. After briefly discussing the economic and technological impact on the USA commercial space industries of the space communications revolution and such large constellation projects, the paper concludes with observations on the trend toward future system architectures using networked groups of much smaller satellites.

  10. Common features in phase-space networks of frustrated spin models and lattice-gas models

    NASA Astrophysics Data System (ADS)

    Wang, Feng; Peng, Yi; Han, Yilong

    2012-02-01

    We mapped the phase spaces of the following four models into networks: (1a) the Ising antiferromagnet on triangular lattice at the ground state and (1b) above the ground state, (2) the six-vertex model (i.e. square ice or spin ice), (3) 1D lattice gas and (4) 2D lattice gas. Their phase-space networks share some common features including the Gaussian degree distribution, the Gaussian spectral density, and the small-world properties. Models 1a, 2 and 3 with long-range correlations in real space exhibit fractal phase spaces, while models 1b and 4 with short-range correlations in real space exhibit non-fractal phase spaces. This result supports one of the untested assumptions in Tsallis's non-extensive statistics.

  11. Distributed audio recording using OFDR with double interrogation

    NASA Astrophysics Data System (ADS)

    Gabai, Haniel; Eyal, Avishay

    2014-05-01

    We introduce a phase sensitive, dynamic and long range fiber-optic sensing system with fully distributed audio recording capabilities. The proposed system implements a recently developed OFDR design, which is based on double interrogation of a sensing fiber with equally-spaced discrete reflectors. In this paper, the ability of each sensing segment to operate as an independent, purely optical audio recorder with little cross-talk artifacts is demonstrated.

  12. The Lowdown on Audio Downloads

    ERIC Educational Resources Information Center

    Farrell, Beth

    2010-01-01

    First offered to public libraries in 2004, downloadable audiobooks have grown by leaps and bounds. According to the Audio Publishers Association, their sales today account for 21% of the spoken-word audio market. It hasn't been easy, however. WMA. DRM. MP3. AAC. File extensions small on letters but very big on consequences for librarians,…

  13. Audio-visual biofeedback for respiratory-gated radiotherapy: Impact of audio instruction and audio-visual biofeedback on respiratory-gated radiotherapy

    SciTech Connect

    George, Rohini; Chung, Theodore D.; Vedam, Sastry S.; Ramakrishnan, Viswanathan; Mohan, Radhe; Weiss, Elisabeth; Keall, Paul J. . E-mail: pjkeall@vcu.edu

    2006-07-01

    Purpose: Respiratory gating is a commercially available technology for reducing the deleterious effects of motion during imaging and treatment. The efficacy of gating is dependent on the reproducibility within and between respiratory cycles during imaging and treatment. The aim of this study was to determine whether audio-visual biofeedback can improve respiratory reproducibility by decreasing residual motion and therefore increasing the accuracy of gated radiotherapy. Methods and Materials: A total of 331 respiratory traces were collected from 24 lung cancer patients. The protocol consisted of five breathing training sessions spaced about a week apart. Within each session the patients initially breathed without any instruction (free breathing), with audio instructions and with audio-visual biofeedback. Residual motion was quantified by the standard deviation of the respiratory signal within the gating window. Results: Audio-visual biofeedback significantly reduced residual motion compared with free breathing and audio instruction. Displacement-based gating has lower residual motion than phase-based gating. Little reduction in residual motion was found for duty cycles less than 30%; for duty cycles above 50% there was a sharp increase in residual motion. Conclusions: The efficiency and reproducibility of gating can be improved by: incorporating audio-visual biofeedback, using a 30-50% duty cycle, gating during exhalation, and using displacement-based gating.

  14. Digital audio and video broadcasting by satellite

    NASA Astrophysics Data System (ADS)

    Yoshino, Takehiko

    In parallel with the progress of the practical use of satellite broadcasting and Hi-Vision or high-definition television technologies, research activities are also in progress to replace the conventional analog broadcasting services with a digital version. What we call 'digitalization' is not a mere technical matter but an important subject which will help promote multichannel or multimedia applications and, accordingly, can change the old concept of mass media, such as television or radio. NHK Science and Technical Research Laboratories has promoted studies of digital bandwidth compression, transmission, and application techniques. The following topics are covered: the trend of digital broadcasting; features of Integrated Services Digital Broadcasting (ISDB); compression encoding and transmission; transmission bit rate in 12 GHz band; number of digital TV transmission channels; multichannel pulse code modulation (PCM) audio broadcasting system via communication satellite; digital Hi-Vision broadcasting; and development of digital audio broadcasting (DAB) for mobile reception in Japan.

  15. Direct broadcast satellite-audio, portable and mobile reception tradeoffs

    NASA Technical Reports Server (NTRS)

    Golshan, Nasser

    1992-01-01

    This paper reports on the findings of a systems tradeoffs study on direct broadcast satellite-radio (DBS-R). Based on emerging advanced subband and transform audio coding systems, four ranges of bit rates: 16-32 kbps, 48-64 kbps, 96-128 kbps and 196-256 kbps are identified for DBS-R. The corresponding grades of audio quality will be subjectively comparable to AM broadcasting, monophonic FM, stereophonic FM, and CD quality audio, respectively. The satellite EIRP's needed for mobile DBS-R reception in suburban areas are sufficient for portable reception in most single family houses when allowance is made for the higher G/T of portable table-top receivers. As an example, the variation of the space segment cost as a function of frequency, audio quality, coverage capacity, and beam size is explored for a typical DBS-R system.

  16. Metrological digital audio reconstruction

    DOEpatents

    Fadeyev; Vitaliy , Haber; Carl

    2004-02-19

    Audio information stored in the undulations of grooves in a medium such as a phonograph record may be reconstructed, with little or no contact, by measuring the groove shape using precision metrology methods coupled with digital image processing and numerical analysis. The effects of damage, wear, and contamination may be compensated, in many cases, through image processing and analysis methods. The speed and data handling capacity of available computing hardware make this approach practical. Two examples used a general purpose optical metrology system to study a 50 year old 78 r.p.m. phonograph record and a commercial confocal scanning probe to study a 1920's celluloid Edison cylinder. Comparisons are presented with stylus playback of the samples and with a digitally re-mastered version of an original magnetic recording. There is also a more extensive implementation of this approach, with dedicated hardware and software.

  17. On equivalent parameter learning in simplified feature space based on Bayesian asymptotic analysis.

    PubMed

    Yamazaki, Keisuke

    2012-07-01

    Parametric models for sequential data, such as hidden Markov models, stochastic context-free grammars, and linear dynamical systems, are widely used in time-series analysis and structural data analysis. Computation of the likelihood function is one of primary considerations in many learning methods. Iterative calculation of the likelihood such as the model selection is still time-consuming though there are effective algorithms based on dynamic programming. The present paper studies parameter learning in a simplified feature space to reduce the computational cost. Simplifying data is a common technique seen in feature selection and dimension reduction though an oversimplified space causes adverse learning results. Therefore, we mathematically investigate a condition of the feature map to have an asymptotically equivalent convergence point of estimated parameters, referred to as the vicarious map. As a demonstration to find vicarious maps, we consider the feature space, which limits the length of data, and derive a necessary length for parameter learning in hidden Markov models. PMID:22503779

  18. The formation method of the feature space for the identification of fatigued bills

    NASA Astrophysics Data System (ADS)

    Kang, Dongshik; Oshiro, Ayumu; Ozawa, Kenji; Mitsui, Ikugo

    2014-10-01

    Fatigued bills make a trouble such as the paper jam in a bill handling machine. In the discrimination of fatigued bills using an acoustic signal, the variation of an observed bill sound is considered to be one of causes in misclassification. Therefore a technique has demanded in order to make the classification of fatigued bills more efficient. In this paper, we proposed the algorithm that extracted feature quantity of bill sound from acoustic signal using the frequency difference, and carried out discrimination experiment of fatigued bill money by Support Vector Machine(SVM). The feature quantity of frequency difference can represent the frequency components of an acoustic signal is varied by the fatigued degree of bill money. The generalization performance of SVM does not depend on the size of dimensions of the feature space, even in a high dimensional feature space such as bill-acoustic signals. Furthermore, SVM can induce an optimal classifier which considers the combination of features by the virtue of polynomial kernel functions.

  19. Segmentation of medical images by feature tracing in a self-dual morphological scale-space

    NASA Astrophysics Data System (ADS)

    Metzler, Volker H.; Thies, Christian; Lehmann, Thomas M.

    2001-07-01

    The multiscale approach derives a segmentation from the evolution of appropriate signal-descriptive features in scale-space. Features that are stable for a wide range of scales are assumed to belong to visually sensible regions. To compensate the well-known drawbacks of linear scale- spaces, the shape-preserving properties of morphological scale-space filtering are utilized. The limiting duality of morphological filters is overcome by a selfdual morphological approach considering both light and dark structures in either the opening or the closing branch of the scale-space. Reconstructive opening/closing-filters enable the scale=analysis of 2D signals, since they are causal with respect to regional maxima/minima. This allows to identify important regions in scale=space via their extrema. Each extremum is assigned a region by a gradient watershed of the corresponding scale. Due to morphological filtering, the scale behavior of the regions is representable by a tree structure describing the spatial inter- and intra-scale relations among regions. The significance of a watershed region is automatically derived from its scale behavior by considering various attributes describing scale-dependent, morphological, and statistical properties of the region. The most significant regions from the segmentation of the image. The algorithm was verified for various medical image domains, such as cytological micrographs, bone x-rays, and cranial NMR slices.

  20. Synthecology: sound use of audio in teleimmersion

    NASA Astrophysics Data System (ADS)

    Baum, Geoffrey; Gotsis, Marientina; Chang, Benjamin; Drinkwater, Robb; St. Clair, Dan

    2006-02-01

    This paper examines historical audio applications used to provide real-time immersive sound for CAVE TM environments and discusses their relative strengths and weaknesses. We examine and explain issues of providing spatialized sound immersion in real-time virtual environments (VEs), some problems with currently used sound servers, and a set of requirements for an 'ideal' sound server. We present the initial configuration of a new cross-platform sound server solution using open source software and the Open Sound Control (OSC) specification for the creation of real-time spatialized audio with CAVE applications, specifically Ygdrasil (Yg) environments. The application, aNother Sound Server (NSS) establishes an application interface (API) using OSC, a logical server layer implemented in Python, and an audio engine using SuperCollider (SC). We discuss spatialization implementation and other features. Finally, we document the Synthecology project which premiered at WIRED NEXTFEST 2005 and was the first VE to use NSS. We also discuss various techniques that enhance presence in networked VEs, as well as possible and planned extensions of NSS.

  1. Study on identifying deciduous forest by the method of feature space transformation

    NASA Astrophysics Data System (ADS)

    Zhang, Xuexia; Wu, Pengfei

    2009-10-01

    The thematic remotely sensed information extraction is always one of puzzling nuts which the remote sensing science faces, so many remote sensing scientists devotes diligently to this domain research. The methods of thematic information extraction include two kinds of the visual interpretation and the computer interpretation, the developing direction of which is intellectualization and comprehensive modularization. The paper tries to develop the intelligent extraction method of feature space transformation for the deciduous forest thematic information extraction in Changping district of Beijing city. The whole Chinese-Brazil resources satellite images received in 2005 are used to extract the deciduous forest coverage area by feature space transformation method and linear spectral decomposing method, and the result from remote sensing is similar to woodland resource census data by Chinese forestry bureau in 2004.

  2. System requirements and design features of Space Station Remote Manipulator System mechanisms

    NASA Technical Reports Server (NTRS)

    Kumar, Rajnish; Hayes, Robert

    1991-01-01

    The Space Station Remote Manipulator System (SSRMS) is a long robotic arm for handling large objects/payloads on the International Space Station Freedom. The mechanical components of the SSRMS include seven joints, two latching end effectors (LEEs), and two boom assemblies. The joints and LEEs are complex aerospace mechanisms. The system requirements and design features of these mechanisms are presented. All seven joints of the SSRMS have identical functional performance. The two LEES are identical. This feature allows either end of the SSRMS to be used as tip or base. As compared to the end effector of the Shuttle Remote Manipulator System, the LEE has a latch and umbilical mechanism in addition to the snare and rigidize mechanisms. The latches increase the interface preload and allow large payloads (up to 116,000 Kg) to be handled. The umbilical connectors provide power, data, and video signal transfer capability to/from the SSRMS.

  3. Matching of flow-cytometry histograms using information theory in feature space.

    PubMed Central

    Zeng, Qing; Wand, Matthew; Young, Alan J.; Rawn, James; Milford, Edgar L.; Mentzer, Steven J.; Greenes, Robert A.

    2002-01-01

    Flow cytometry is a widely available technique for analyzing cell-surface protein expression. Data obtained from flow cytometry is frequently used to produce fluorescence intensity histograms. Comparison of histograms can be useful in the identification of unknown molecules and in the analysis of protein expression. In this study, we examined the combination of a new smoothing technique called SiZer with information theory to measure the difference between cytometry histograms. SiZer provides cross-bandwidth smoothing and allowed analysis in feature space. The new methods were tested on a panel of monoclonal antibodies raised against proteins expressed on peripheral blood lymphocytes and compared with previous methods. The findings suggest that comparing information content of histograms in feature space is effective and efficient for identifying antibodies with similar cell-surface binding patterns. PMID:12463961

  4. Do infants encode feature and geometry information in a two-dimensional space?

    PubMed

    Lábadi, Beatrix; Horváth, Diana Á; Palotai, Róbert

    2012-06-01

    Geometric form perception and its role in reorienting process have been studied extensively in children, but little is known about its early origin in infancy. Here, we present findings of three experiments that used looking-time paradigm to test infants' sensitivity to geometric and feature information in two-dimensional visual display. One-year-old infants participated in spatial search tasks, which were manipulated regarding the display movement (static, visible rotation and invisible rotation) as a degree of disorientation. The results showed that infants were able to create expectation about a hiding location based on the geometry of a rectangle only in the directionally stable search space, whereas they were capable to use feature cues (color) for reorienting even in that condition when the spatial display was rotated and they were allowed to track the display motion. However, infants did not use either geometry or feature properties of 2-D space in an orientation invariant manner. The findings are discussed within the theory of reorientation with respect to the 2-dimensional space. PMID:22721741

  5. Radiometric responsivity determination for Feature Identification and Location Experiment (FILE) flown on space shuttle mission

    NASA Technical Reports Server (NTRS)

    Wilson, R. G.; Davis, R. E.; Wright, R. E., Jr.; Sivertson, W. E., Jr.; Bullock, G. F.

    1986-01-01

    A procedure was developed to obtain the radiometric (radiance) responsivity of the Feature Identification and Local Experiment (FILE) instrument in preparation for its flight on Space Shuttle Mission 41-G (November 1984). This instrument was designed to obtain Earth feature radiance data in spectral bands centered at 0.65 and 0.85 microns, along with corroborative color and color-infrared photographs, and to collect data to evaluate a technique for in-orbit autonomous classification of the Earth's primary features. The calibration process incorporated both solar radiance measurements and radiative transfer model predictions in estimating expected radiance inputs to the FILE on the Shuttle. The measured data are compared with the model predictions, and the differences observed are discussed. Application of the calibration procedure to the FILE over an 18-month period indicated a constant responsivity characteristic. This report documents the calibration procedure and the associated radiometric measurements and predictions that were part of the instrument preparation for flight.

  6. A centralized audio presentation manager

    SciTech Connect

    Papp, A.L. III; Blattner, M.M.

    1994-05-16

    The centralized audio presentation manager addresses the problems which occur when multiple programs running simultaneously attempt to use the audio output of a computer system. Time dependence of sound means that certain auditory messages must be scheduled simultaneously, which can lead to perceptual problems due to psychoacoustic phenomena. Furthermore, the combination of speech and nonspeech audio is examined; each presents its own problems of perceptibility in an acoustic environment composed of multiple auditory streams. The centralized audio presentation manager receives abstract parameterized message requests from the currently running programs, and attempts to create and present a sonic representation in the most perceptible manner through the use of a theoretically and empirically designed rule set.

  7. The Audio-Tutorial System

    ERIC Educational Resources Information Center

    Postlethwait, S. N.

    1970-01-01

    Describes the audio-tutorial program in Botany at Purdue University. Advantages include adaptability to individual stduent needs, integration of laboratory activities and information giving, aid flexibility in use of media and means of presentation. (EB)

  8. Large-scale earth surface thermal radiative features in space observation

    NASA Astrophysics Data System (ADS)

    Yang, Fan; Han, Yuge; Xuan, Yimin

    2015-08-01

    It is necessary to complete the earth thermal radiative modeling, since it is the most important background in space infrared observation. A new method was proposed to calculate the earth thermal infrared radiation combined with remote sensing technology. The simplified model also was proposed when the solar radiative impact is neglected properly. The practical split-window algorithm was used to retrieve the global surface temperature from MODIS data products. Integrated with MODTRAN code to calculate the atmospheric radiation and transmittance, the earth thermal infrared features were calculated in typical months. Moreover the radiance dependence on viewing angle was discussed. Through the comparison with CERES measurement results, this model has been proved effective and practicable, and that it would have a further application in space thermal environment analysis or space infrared observation technology.

  9. The nature and origin of periodically spaced tectonic features on Mars

    NASA Technical Reports Server (NTRS)

    Watters, Thomas R.

    1991-01-01

    The final report on the nature and origin of periodically spaced tectonic features on Mars is presented. The focus of this investigation was to characterize and model the periodically spaced wrinkle ridges observed in ridged plains material on Mars. The investigation centered on the wrinkle ridges in ridged plains material on the Tharsis Plateau. Wrinkle ridges are interpreted to be structural in origin, resulting from buckling followed by reverse or thrust faulting of the ridged plains material. The study extended beyond Tharsis to other ridged plains units, particularly those in Hesperia Planum. As a corollary, an analysis of the spacing of the anticlinal ridges of the Yakima Fold Belt of the Columbia Plateau in the NW United States was undertaken.

  10. Audio-Visual Temporal Recalibration Can be Constrained by Content Cues Regardless of Spatial Overlap

    PubMed Central

    Roseboom, Warrick; Kawabe, Takahiro; Nishida, Shin’Ya

    2013-01-01

    It has now been well established that the point of subjective synchrony for audio and visual events can be shifted following exposure to asynchronous audio-visual presentations, an effect often referred to as temporal recalibration. Recently it was further demonstrated that it is possible to concurrently maintain two such recalibrated estimates of audio-visual temporal synchrony. However, it remains unclear precisely what defines a given audio-visual pair such that it is possible to maintain a temporal relationship distinct from other pairs. It has been suggested that spatial separation of the different audio-visual pairs is necessary to achieve multiple distinct audio-visual synchrony estimates. Here we investigated if this is necessarily true. Specifically, we examined whether it is possible to obtain two distinct temporal recalibrations for stimuli that differed only in featural content. Using both complex (audio visual speech; see Experiment 1) and simple stimuli (high and low pitch audio matched with either vertically or horizontally oriented Gabors; see Experiment 2) we found concurrent, and opposite, recalibrations despite there being no spatial difference in presentation location at any point throughout the experiment. This result supports the notion that the content of an audio-visual pair alone can be used to constrain distinct audio-visual synchrony estimates regardless of spatial overlap. PMID:23658549

  11. The SWRL Audio Laboratory System (ALS): An Integrated Configuration for Psychomusicology Research. Technical Report 51.

    ERIC Educational Resources Information Center

    Williams, David Brian; Hoskin, Richard K.

    This report describes features of the Audio Laboratory System (ALS), a device which supports research activities of the Southwest Regional Laboratory's Music Program. The ALS is used primarily to generate recorded audio tapes for psychomusicology research related to children's perception and learning of music concepts such as pitch, loudness,…

  12. Non-RF wireless helmet-mounted display and two-way audio connectivity using covert free-space optical communications

    NASA Astrophysics Data System (ADS)

    Strauss, M.; Volfson, L.

    2011-06-01

    Providing the warfighter with Head or Helmet Mounted Displays (HMDs) while in tracked vehicles provides a means to visually maintain access to systems information while in a high vibration environment. The high vibration and unique environment of military tracked and turreted vehicles impact the ability to distinctly see certain information on an HMD, especially small font size or graphics and information that requires long fixation (staring), rather than a brief or peripheral glance. The military and commercial use of HMDs was compiled from market research, market trends, and user feedback. Lessons learned from previous military and commercial use of HMD products were derived to determine the feasibility of HMDs use in the high vibration and the unique environments of tracked vehicles. The results are summarized into factors that determine HMD features which must be specified for successful implementation.

  13. Nanoscale Analysis of Space-Weathering Features in Soils from Itokawa

    NASA Technical Reports Server (NTRS)

    Thompson, M. S.; Christoffersen, R.; Zega, T. J.; Keller, L. P.

    2014-01-01

    Space weathering alters the spectral properties of airless body surface materials by redden-ing and darkening their spectra and attenuating characteristic absorption bands, making it challenging to characterize them remotely [1,2]. It also causes a discrepency between laboratory analysis of meteorites and remotely sensed spectra from asteroids, making it difficult to associate meteorites with their parent bodies. The mechanisms driving space weathering include mi-crometeorite impacts and the interaction of surface materials with solar energetic ions, particularly the solar wind. These processes continuously alter the microchemical and structural characteristics of exposed grains on airless bodies. The change of these properties is caused predominantly by the vapor deposition of reduced Fe and FeS nanoparticles (npFe(sup 0) and npFeS respectively) onto the rims of surface grains [3]. Sample-based analysis of space weathering has tra-ditionally been limited to lunar soils and select asteroidal and lunar regolith breccias [3-5]. With the return of samples from the Hayabusa mission to asteroid Itoka-wa [6], for the first time we are able to compare space-weathering features on returned surface soils from a known asteroidal body. Analysis of these samples will contribute to a more comprehensive model for how space weathering varies across the inner solar system. Here we report detailed microchemical and microstructal analysis of surface grains from Itokawa.

  14. Prediction of protein structure classes using hybrid space of multi-profile Bayes and bi-gram probability feature spaces.

    PubMed

    Hayat, Maqsood; Tahir, Muhammad; Khan, Sher Afzal

    2014-04-01

    Proteins are the executants of biological functions in living organisms. Comprehension of protein structure is a challenging problem in the era of proteomics, computational biology, and bioinformatics because of its pivotal role in protein folding patterns. Owing to the large exploration of protein sequences in protein databanks and intricacy of protein structures, experimental and theoretical methods are insufficient for prediction of protein structure classes. Therefore, it is highly desirable to develop an accurate, reliable, and high throughput computational model to predict protein structure classes correctly from polygenetic sequences. In this regard, we propose a promising model employing hybrid descriptor space in conjunction with optimized evidence-theoretic K-nearest neighbor algorithm. Hybrid space is the composition of two descriptor spaces including Multi-profile Bayes and bi-gram probability. In order to enhance the generalization power of the classifier, we have selected high discriminative descriptors from the hybrid space using particle swarm optimization, a well-known evolutionary feature selection technique. Performance evaluation of the proposed model is performed using the jackknife test on three low similarity benchmark datasets including 25PDB, 1189, and 640. The success rates of the proposed model are 87.0%, 86.6%, and 88.4%, respectively on the three benchmark datasets. The comparative analysis exhibits that our proposed model has yielded promising results compared to the existing methods in the literature. In addition, our proposed prediction system might be helpful in future research particularly in cases where the major focus of research is on low similarity datasets. PMID:24384128

  15. Assessing efficiency of spatial sampling using combined coverage analysis in geographical and feature spaces

    NASA Astrophysics Data System (ADS)

    Hengl, Tomislav

    2015-04-01

    Efficiency of spatial sampling largely determines success of model building. This is especially important for geostatistical mapping where an initial sampling plan should provide a good representation or coverage of both geographical (defined by the study area mask map) and feature space (defined by the multi-dimensional covariates). Otherwise the model will need to extrapolate and, hence, the overall uncertainty of the predictions will be high. In many cases, geostatisticians use point data sets which are produced using unknown or inconsistent sampling algorithms. Many point data sets in environmental sciences suffer from spatial clustering and systematic omission of feature space. But how to quantify these 'representation' problems and how to incorporate this knowledge into model building? The author has developed a generic function called 'spsample.prob' (Global Soil Information Facilities package for R) and which simultaneously determines (effective) inclusion probabilities as an average between the kernel density estimation (geographical spreading of points; analysed using the spatstat package in R) and MaxEnt analysis (feature space spreading of points; analysed using the MaxEnt software used primarily for species distribution modelling). The output 'iprob' map indicates whether the sampling plan has systematically missed some important locations and/or features, and can also be used as an input for geostatistical modelling e.g. as a weight map for geostatistical model fitting. The spsample.prob function can also be used in combination with the accessibility analysis (cost of field survey are usually function of distance from the road network, slope and land cover) to allow for simultaneous maximization of average inclusion probabilities and minimization of total survey costs. The author postulates that, by estimating effective inclusion probabilities using combined geographical and feature space analysis, and by comparing survey costs to representation

  16. Machine learning approaches for discrimination of Extracellular Matrix proteins using hybrid feature space.

    PubMed

    Ali, Farman; Hayat, Maqsood

    2016-08-21

    Extracellular Matrix (ECM) proteins are the vital type of proteins that are secreted by resident cells. ECM proteins perform several significant functions including adhesion, differentiation, cell migration and proliferation. In addition, ECM proteins regulate angiogenesis process, embryonic development, tumor growth and gene expression. Due to tremendous biological significance of the ECM proteins and rapidly increases of protein sequences in databases, it is indispensable to introduce a new high throughput computation model that can accurately identify ECM proteins. Various traditional models have been developed, but they are laborious and tedious. In this work, an effective and high throughput computational classification model is proposed for discrimination of ECM proteins. In this model, protein sequences are formulated using amino acid composition, pseudo amino acid composition (PseAAC) and di-peptide composition (DPC) techniques. Further, various combination of feature extraction techniques are fused to form hybrid feature spaces. Several classifiers were employed. Among these classifiers, K-Nearest Neighbor obtained outstanding performance in combination with the hybrid feature space of PseAAC and DPC. The obtained accuracy of our proposed model is 96.76%, which the highest success rate has been reported in the literature so far. PMID:27179459

  17. EEMD Independent Extraction for Mixing Features of Rotating Machinery Reconstructed in Phase Space

    PubMed Central

    Ma, Zaichao; Wen, Guangrui; Jiang, Cheng

    2015-01-01

    Empirical Mode Decomposition (EMD), due to its adaptive decomposition property for the non-linear and non-stationary signals, has been widely used in vibration analyses for rotating machinery. However, EMD suffers from mode mixing, which is difficult to extract features independently. Although the improved EMD, well known as the ensemble EMD (EEMD), has been proposed, mode mixing is alleviated only to a certain degree. Moreover, EEMD needs to determine the amplitude of added noise. In this paper, we propose Phase Space Ensemble Empirical Mode Decomposition (PSEEMD) integrating Phase Space Reconstruction (PSR) and Manifold Learning (ML) for modifying EEMD. We also provide the principle and detailed procedure of PSEEMD, and the analyses on a simulation signal and an actual vibration signal derived from a rubbing rotor are performed. The results show that PSEEMD is more efficient and convenient than EEMD in extracting the mixing features from the investigated signal and in optimizing the amplitude of the necessary added noise. Additionally PSEEMD can extract the weak features interfered with a certain amount of noise. PMID:25871723

  18. Development of sensitivity to spacing versus feature changes in pictures of houses: Evidence for slow development of a general spacing detection mechanism?

    PubMed

    Robbins, Rachel A; Shergill, Yaadwinder; Maurer, Daphne; Lewis, Terri L

    2011-07-01

    Adults are expert at recognizing faces, in part because of exquisite sensitivity to the spacing of facial features. Children are poorer than adults at recognizing facial identity and less sensitive to spacing differences. Here we examined the specificity of the immaturity by comparing the ability of 8-year-olds, 14-year-olds, and adults to discriminate houses differing in the spacing between features versus those differing in the shape of the features themselves. By 8 years of age, children were more accurate for discriminations involving the feature set compared with the spacing set, and the difference in accuracy compared with adults was greater for the spacing set than for the feature set. Importantly, when sets were matched in difficulty for adults, this greater immaturity on the spacing set than on the feature set remained. The results suggest that, at least by age 8, immaturities in sensitivity to the spacing of features may be related to immaturities in general perceptual mechanisms rather than face-specific mechanisms. PMID:21397248

  19. Specific features of space-time variations of ozone during the development of intensive tropical disturbances

    NASA Technical Reports Server (NTRS)

    Nerushev, Alexander F.; Vasiliev, Victor I.

    1994-01-01

    An analysis of specific features of space-time variations of ozone in the tropical areas which has been performed on the basis of processing of the results of special expedition studies in the Atlantic and Pacific in 1987-1990 and the data of observations at the stations of the world ozonometric network over the 25-year period. The existence of a cause-and-effect relation has been revealed between the processes determining tropical cyclone (TC) development, and specific features of variations of the total content of ozone (TCO) and the vertical distribution of ozone (VDO) in the regions of TC action. Characteristic features of day-to-day and daily variations of TCO during TC development have been found. On the periphery of a developing TC, 1-4 days before it reaches the stage of storm, TCO increases, on average, by 5-8 percent, and a substantial increase in the concentration of ozone occurs in the middle and upper troposphere. The most probable physical mechanisms relating the observed specific features of ozone variations to TC evolution have been suggested. A hypothesis of the possibility of using ozone as an indicator for early prediction of TC development has been substantiated.

  20. Optimizing view/illumination geometry for terrestrial features using Space Shuttle and aerial polarimetry

    NASA Technical Reports Server (NTRS)

    Israel, Steven A.; Holly, Mark H.; Whitehead, Victor S.

    1992-01-01

    This paper describes to relationship of polarimetric observations from orbital and aerial platforms and the determination optimum sun-target-sensor geometry. Polarimetric observations were evaluated for feature discrimination. The Space Shuttle experiment was performed using two boresighted Hasselblad 70 mm cameras with identical settings with linear polarizing filters aligned orthogonally about the optic axis. The aerial experiment was performed using a single 35 mm Nikon FE2 and rotating the linear polarizing filter 90 deg to acquire both minimum and maximum photographs. Characteristic curves were created by covertype and waveband for both aerial and Space Shuttle imagery. Though significant differences existed between the two datasets, the observed polarimetric signatures were unique and separable.

  1. Features of the solar array drive mechanism for the space telescope

    NASA Technical Reports Server (NTRS)

    Hostenkamp, R. G.

    1985-01-01

    The solar array drive mechanism for the Space Telescope embodies several features not customarily found on solar array drives. Power and signal transfer is achieved by means of a flexible wire harness for which the chosen solution, consisting of 168 standard wires, is described. The torque performance data of the harness over its temperature range are presented. The off load system which protects the bearings from the launch loads is released by a trigger made from Nitinol, the memory alloy. The benefits of memory alloy and the caveats for the design are briefly discussed. The design of the off load system is described and test experience is reported.

  2. Supervised pixel classification using a feature space derived from an artificial visual system

    NASA Technical Reports Server (NTRS)

    Baxter, Lisa C.; Coggins, James M.

    1991-01-01

    Image segmentation involves labelling pixels according to their membership in image regions. This requires the understanding of what a region is. Using supervised pixel classification, the paper investigates how groups of pixels labelled manually according to perceived image semantics map onto the feature space created by an Artificial Visual System. Multiscale structure of regions are investigated and it is shown that pixels form clusters based on their geometric roles in the image intensity function, not by image semantics. A tentative abstract definition of a 'region' is proposed based on this behavior.

  3. On pose determination using point features. [vision system for space robots

    NASA Technical Reports Server (NTRS)

    Hwang, Vincent; Keizer, Richard; Winkert, Tom; Spidaliere, Peter

    1992-01-01

    Consideration is given to the vision subsystem of an Orbital Replacement Unit (ORU) that was placed onto its base at Goddard Space Flight Center by a PUMA 762 robot equipped with a wrist-mounted CCD camera and a wrist-mounted force sensor. It is found that a simple adaptive thresholding method works quite well for images taken under various lighting conditions. The pose computed using the quadrangle method is reasonable for real images. In the presence of image feature noise the accuracy of the computed pose can be considerably reduced. This problem can be solved by using a 3D marker and an alternative pose computation algorithm.

  4. A systematic comparison of feature space effects on disease classifier performance for phenotype identification of five diseases.

    PubMed

    Kotfila, Christopher; Uzuner, Özlem

    2015-12-01

    Automated phenotype identification plays a critical role in cohort selection and bioinformatics data mining. Natural Language Processing (NLP)-informed classification techniques can robustly identify phenotypes in unstructured medical notes. In this paper, we systematically assess the effect of naive, lexically normalized, and semantic feature spaces on classifier performance for obesity, atherosclerotic cardiovascular disease (CAD), hyperlipidemia, hypertension, and diabetes. We train support vector machines (SVMs) using individual feature spaces as well as combinations of these feature spaces on two small training corpora (730 and 790 documents) and a combined (1520 documents) training corpus. We assess the importance of feature spaces and training data size on SVM model performance. We show that inclusion of semantically-informed features does not statistically improve performance for these models. The addition of training data has weak effects of mixed statistical significance across disease classes suggesting larger corpora are not necessary to achieve relatively high performance with these models. PMID:26241355

  5. Audio-Visual Training in Children with Reading Disabilities

    ERIC Educational Resources Information Center

    Magnan, Annie; Ecalle, Jean

    2006-01-01

    This study tested the effectiveness of audio-visual training in the discrimination of the phonetic feature of voicing on the recognition of written words by young children deemed to at risk of dyslexia (experiment 1) as well as on dyslexic children's phonological skills (experiment 2). In addition, the third experiment studied the effectiveness of…

  6. Using MPEG-7 audio descriptors for music querying

    NASA Astrophysics Data System (ADS)

    Gruhne, M.; Dittmar, C.

    2006-08-01

    Due to the growing amount of digital audio an increasing need to automatically categorize music and to create self-controlled and suitable playlists has been emerged. A few approaches to this task relying on low-level features have been published so far. Unfortunately the results utilizing those technologies are not sufficient yet. This paper gives an introduction how to enhance the results with regard to the perceptual similarity using different high-level descriptors and a powerful interaction between the algorithm and the user to consider his preferences. A successful interaction between server and client requires a powerful standardized query language. This paper describes the tools of the MPEG-7 Audio standard in detail and gives examples of already established query languages. Furthermore the requirements of a multimedia query language are identified and its application is exemplified by an automatic audio creation system using a query language.

  7. Audio signal recognition for speech, music, and environmental sounds

    NASA Astrophysics Data System (ADS)

    Ellis, Daniel P. W.

    2003-10-01

    Human listeners are very good at all kinds of sound detection and identification tasks, from understanding heavily accented speech to noticing a ringing phone underneath music playing at full blast. Efforts to duplicate these abilities on computer have been particularly intense in the area of speech recognition, and it is instructive to review which approaches have proved most powerful, and which major problems still remain. The features and models developed for speech have found applications in other audio recognition tasks, including musical signal analysis, and the problems of analyzing the general ``ambient'' audio that might be encountered by an auditorily endowed robot. This talk will briefly review statistical pattern recognition for audio signals, giving examples in several of these domains. Particular emphasis will be given to common aspects and lessons learned.

  8. Note-accurate audio segmentation based on MPEG-7

    NASA Astrophysics Data System (ADS)

    Wellhausen, Jens

    2003-12-01

    Segmenting audio data into the smallest musical components is the basis for many further meta data extraction algorithms. For example, an automatic music transcription system needs to know where the exact boundaries of each tone are. In this paper a note accurate audio segmentation algorithm based on MPEG-7 low level descriptors is introduced. For a reliable detection of different notes, both features in the time and the frequency domain are used. Because of this, polyphonic instrument mixes and even melodies characterized by human voices can be examined with this alogrithm. For testing and verification of the note accurate segmentation, a simple music transcription system was implemented. The dominant frequency within each segment is used to build a MIDI file representing the processed audio data.

  9. Audio-guided audiovisual data segmentation, indexing, and retrieval

    NASA Astrophysics Data System (ADS)

    Zhang, Tong; Kuo, C.-C. Jay

    1998-12-01

    While current approaches for video segmentation and indexing are mostly focused on visual information, audio signals may actually play a primary role in video content parsing. In this paper, we present an approach for automatic segmentation, indexing, and retrieval of audiovisual data, based on audio content analysis. The accompanying audio signal of audiovisual data is first segmented and classified into basic types, i.e., speech, music, environmental sound, and silence. This coarse-level segmentation and indexing step is based upon morphological and statistical analysis of several short-term features of the audio signals. Then, environmental sounds are classified into finer classes, such as applause, explosions, bird sounds, etc. This fine-level classification and indexing step is based upon time- frequency analysis of audio signals and the use of the hidden Markov model as the classifier. On top of this archiving scheme, an audiovisual data retrieval system is proposed. Experimental results show that the proposed approach has an accuracy rate higher than 90 percent for the coarse-level classification, and higher than 85 percent for the fine-level classification. Examples of audiovisual data segmentation and retrieval are also provided.

  10. Virtual environment display for a 3D audio room simulation

    NASA Technical Reports Server (NTRS)

    Chapin, William L.; Foster, Scott H.

    1992-01-01

    The development of a virtual environment simulation system integrating a 3D acoustic audio model with an immersive 3D visual scene is discussed. The system complements the acoustic model and is specified to: allow the listener to freely move about the space, a room of manipulable size, shape, and audio character, while interactively relocating the sound sources; reinforce the listener's feeling of telepresence in the acoustical environment with visual and proprioceptive sensations; enhance the audio with the graphic and interactive components, rather than overwhelm or reduce it; and serve as a research testbed and technology transfer demonstration. The hardware/software design of two demonstration systems, one installed and one portable, are discussed through the development of four iterative configurations.

  11. QRDA: Quantum Representation of Digital Audio

    NASA Astrophysics Data System (ADS)

    Wang, Jian

    2016-03-01

    Multimedia refers to content that uses a combination of different content forms. It includes two main medias: image and audio. However, by contrast with the rapid development of quantum image processing, quantum audio almost never been studied. In order to change this status, a quantum representation of digital audio (QRDA) is proposed in this paper to present quantum audio. QRDA uses two entangled qubit sequences to store the audio amplitude and time information. The two qubit sequences are both in basis state: |0> and |1>. The QRDA audio preparation from initial state |0> is given to store an audio in quantum computers. Then some exemplary quantum audio processing operations are performed to indicate QRDA's usability.

  12. Digital Audio Application to Short Wave Broadcasting

    NASA Technical Reports Server (NTRS)

    Chen, Edward Y.

    1997-01-01

    Digital audio is becoming prevalent not only in consumer electornics, but also in different broadcasting media. Terrestrial analog audio broadcasting in the AM and FM bands will be eventually be replaced by digital systems.

  13. Audio direct broadcast satellites

    NASA Technical Reports Server (NTRS)

    Miller, J. E.

    1983-01-01

    Satellite sound broadcasting is, as the name implies, the use of satellite techniques and technology to broadcast directly from space to low-cost, consumer-quality receivers the types of sound programs commonly received in the AM and FM broadcast bands. It would be a ubiquitous service available to the general public in the home, in the car, and out in the open.

  14. Development of Sensitivity to Spacing Versus Feature Changes in Pictures of Houses: Evidence for Slow Development of a General Spacing Detection Mechanism?

    ERIC Educational Resources Information Center

    Robbins, Rachel A.; Shergill, Yaadwinder; Maurer, Daphne; Lewis, Terri L.

    2011-01-01

    Adults are expert at recognizing faces, in part because of exquisite sensitivity to the spacing of facial features. Children are poorer than adults at recognizing facial identity and less sensitive to spacing differences. Here we examined the specificity of the immaturity by comparing the ability of 8-year-olds, 14-year-olds, and adults to…

  15. Radioactive Decay: Audio Data Collection

    ERIC Educational Resources Information Center

    Struthers, Allan

    2009-01-01

    Many phenomena generate interesting audible time series. This data can be collected and processed using audio software. The free software package "Audacity" is used to demonstrate the process by recording, processing, and extracting click times from an inexpensive radiation detector. The high quality of the data is demonstrated with a simple…

  16. Audio-Visual Teaching Machines.

    ERIC Educational Resources Information Center

    Dorsett, Loyd G.

    An audiovisual teaching machine (AVTM) presents programed audio and visual material simultaneously to a student and accepts his response. If his response is correct, the machine proceeds with the lesson; if it is incorrect, the machine so indicates and permits another choice (linear) or automatically presents supplementary material (branching).…

  17. Engaging Students with Audio Feedback

    ERIC Educational Resources Information Center

    Cann, Alan

    2014-01-01

    Students express widespread dissatisfaction with academic feedback. Teaching staff perceive a frequent lack of student engagement with written feedback, much of which goes uncollected or unread. Published evidence shows that audio feedback is highly acceptable to students but is underused. This paper explores methods to produce and deliver audio…

  18. Audio-Visual Materials Catalog.

    ERIC Educational Resources Information Center

    Anderson (M.D.) Hospital and Tumor Inst., Houston, TX.

    This catalog lists 27 audiovisual programs produced by the Department of Medical Communications of the University of Texas M. D. Anderson Hospital and Tumor Institute for public distribution. Video tapes, 16 mm. motion pictures and slide/audio series are presented dealing mostly with cancer and related subjects. The programs are intended for…

  19. Audio/ Videoconferencing Packages: Low Cost

    ERIC Educational Resources Information Center

    Treblay, Remy; Fyvie, Barb; Koritko, Brenda

    2005-01-01

    A comparison was conducted of "Voxwire MeetingRoom" and "iVocalize" v4.1.0.3, both Web-conferencing products using voice-over-Internet protocol (VoIP) to provide unlimited, inexpensive, international audio communication, and high-quality Web-conferencing fostering collaborative learning. The study used the evaluation criteria used in earlier…

  20. A Simple Audio Conductivity Device.

    ERIC Educational Resources Information Center

    Berenato, Gregory; Maynard, David F.

    1997-01-01

    Describes a simple audio conductivity device built to address the problem of the lack of sensitivity needed to measure small differences in conductivity in crude conductivity devices. Uses a 9-V battery as a power supply and allows the relative resistance differences between substances to be detected by the frequency of its audible tones. Presents…

  1. Audio Frequency Analysis in Mobile Phones

    ERIC Educational Resources Information Center

    Aguilar, Horacio Munguía

    2016-01-01

    A new experiment using mobile phones is proposed in which its audio frequency response is analyzed using the audio port for inputting external signal and getting a measurable output. This experiment shows how the limited audio bandwidth used in mobile telephony is the main cause of the poor speech quality in this service. A brief discussion is…

  2. Audio-Visual Aids: Historians in Blunderland.

    ERIC Educational Resources Information Center

    Decarie, Graeme

    1988-01-01

    A history professor relates his experiences producing and using audio-visual material and warns teachers not to rely on audio-visual aids for classroom presentations. Includes examples of popular audio-visual aids on Canada that communicate unintended, inaccurate, or unclear ideas. Urges teachers to exercise caution in the selection and use of…

  3. 50 CFR 27.72 - Audio equipment.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 50 Wildlife and Fisheries 6 2010-10-01 2010-10-01 false Audio equipment. 27.72 Section 27.72 Wildlife and Fisheries UNITED STATES FISH AND WILDLIFE SERVICE, DEPARTMENT OF THE INTERIOR (CONTINUED) THE... Audio equipment. The operation or use of audio devices including radios, recording and playback...

  4. 36 CFR 1002.12 - Audio disturbances.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... 36 Parks, Forests, and Public Property 3 2010-07-01 2010-07-01 false Audio disturbances. 1002.12... RECREATION § 1002.12 Audio disturbances. (a) The following are prohibited: (1) Operating motorized equipment or machinery such as an electric generating plant, motor vehicle, motorized toy, or an audio...

  5. Joint Spatial-Spectral Feature Space Clustering for Speech Activity Detection from ECoG Signals

    PubMed Central

    Kanas, Vasileios G.; Mporas, Iosif; Benz, Heather L.; Sgarbas, Kyriakos N.; Bezerianos, Anastasios; Crone, Nathan E.

    2014-01-01

    Brain machine interfaces for speech restoration have been extensively studied for more than two decades. The success of such a system will depend in part on selecting the best brain recording sites and signal features corresponding to speech production. The purpose of this study was to detect speech activity automatically from electrocorticographic signals based on joint spatial-frequency clustering of the ECoG feature space. For this study, the ECoG signals were recorded while a subject performed two different syllable repetition tasks. We found that the optimal frequency resolution to detect speech activity from ECoG signals was 8 Hz, achieving 98.8% accuracy by employing support vector machines (SVM) as a classifier. We also defined the cortical areas that held the most information about the discrimination of speech and non-speech time intervals. Additionally, the results shed light on the distinct cortical areas associated with the two syllable repetition tasks and may contribute to the development of portable ECoG-based communication. PMID:24658248

  6. Interactive Feature Space Explorer© for Multi–Modal Magnetic Resonance Imaging

    PubMed Central

    Türkbey, Barış; Choyke, Peter L.; Akin, Oguz; Aras, Ömer; Mun, Seong K.

    2015-01-01

    Wider information content of multi–modal biomedical imaging is advantageous for detection, diagnosis and prognosis of various pathologies. However, the necessity to evaluate a large number images might hinder these advantages and reduce the efficiency. Herein, a new computer aided approach based on the utilization of feature space (FS) with reduced reliance on multiple image evaluations is proposed for research and routine clinical use. The method introduces the physician experience into the discovery process of FS biomarkers for addressing biological complexity, e.g., disease heterogeneity. This, in turn, elucidates relevant biophysical information which would not be available when automated algorithms are utilized. Accordingly, the prototype platform was designed and built for interactively investigating the features and their corresponding anatomic loci in order to identify pathologic FS regions. While the platform might be potentially beneficial in decision support generally and specifically for evaluating outlier cases, it is also potentially suitable for accurate ground truth determination in FS for algorithm development. Initial assessments conducted on two different pathologies from two different institutions provided valuable biophysical perspective. Investigations of the prostate magnetic resonance imaging data resulted in locating a potential aggressiveness biomarker in prostate cancer. Preliminary findings on renal cell carcinoma imaging data demonstrated potential for characterization of disease subtypes in the FS. PMID:25868623

  7. Learning one-to-many mapping functions for audio-visual integrated perception

    NASA Astrophysics Data System (ADS)

    Lim, Jung-Hui; Oh, Do-Kwan; Lee, Soo-Young

    2010-04-01

    In noisy environment the human speech perception utilizes visual lip-reading as well as audio phonetic classification. This audio-visual integration may be done by combining the two sensory features at the early stage. Also, the top-down attention may integrate the two modalities. For the sensory feature fusion we introduce mapping functions between the audio and visual manifolds. Especially, we present an algorithm to provide one-to-many mapping function for the videoto- audio mapping. The top-down attention is also presented to integrate both the sensory features and classification results of both modalities, which is able to explain McGurk effect. Each classifier is separately implemented by the Hidden-Markov Model (HMM), but the two classifiers are combined at the top level and interact by the top-down attention.

  8. Two-dimensional audio watermark for MPEG AAC audio

    NASA Astrophysics Data System (ADS)

    Tachibana, Ryuki

    2004-06-01

    Since digital music is often stored in a compressed file, it is desirable that an audio watermarking method in a content management system handles compressed files. Using an audio watermarking method that directly manipulates compressed files makes it unnecessary to decompress the files before embedding or detection, so more files can be processed per unit time. However, it is difficult to detect a watermark in a compressed file that has been compressed after the file was watermarked. This paper proposes an MPEG Advanced Audio Coding (AAC) bitstream watermarking method using a two-dimensional pseudo-random array. Detection is done by correlating the absolute values of the recovered MDCT coefficients and the pseudo-random array. Since the embedding algorithm uses the same pseudo-random values for two adjacent overlapping frames and the detection algorithm selects the better frame in the two by comparing detected watermark strengths, it is possible to detect a watermark from a compressed file that was compressed after the watermark was embedded in the original uncompressed file. Though the watermark is not detected as clearly in this case, the watermark can still be detected even when the watermark was embedded in a compressed file and the file was then decompressed, trimmed, and compressed again.

  9. Audio-visual interactions in environment assessment.

    PubMed

    Preis, Anna; Kociński, Jędrzej; Hafke-Dys, Honorata; Wrzosek, Małgorzata

    2015-08-01

    The aim of the study was to examine how visual and audio information influences audio-visual environment assessment. Original audio-visual recordings were made at seven different places in the city of Poznań. Participants of the psychophysical experiments were asked to rate, on a numerical standardized scale, the degree of comfort they would feel if they were in such an environment. The assessments of audio-visual comfort were carried out in a laboratory in four different conditions: (a) audio samples only, (b) original audio-visual samples, (c) video samples only, and (d) mixed audio-visual samples. The general results of this experiment showed a significant difference between the investigated conditions, but not for all the investigated samples. There was a significant improvement in comfort assessment when visual information was added (in only three out of 7 cases), when conditions (a) and (b) were compared. On the other hand, the results show that the comfort assessment of audio-visual samples could be changed by manipulating the audio rather than the video part of the audio-visual sample. Finally, it seems, that people could differentiate audio-visual representations of a given place in the environment based rather of on the sound sources' compositions than on the sound level. Object identification is responsible for both landscape and soundscape grouping. PMID:25863510

  10. Characterization of Clastic Dikes Using Controlled Source Audio Magnetotellurics

    NASA Astrophysics Data System (ADS)

    Persichetti, J. A.; Alumbaugh, D.

    2001-12-01

    A site consisting of 3D geology on the Hanford Reservation in Hanford, Washington, has been surveyed using Controlled Source Audio Magnetotellurics (CSAMT) to determine the method's ability to detect clastic dikes. The dikes are fine-grained, soft-sediment intrusions, formed by the buoyant rise of buried, unconsolidated, water rich mud into overlying unconsolidated sediment. The dikes are of major importance because they may act as natural barriers inhibiting the spread of contaminants, or as conduits, allowing the contaminants to be quickly wicked away from the contaminant storage tanks that may be located in close vicinity of the dikes. The field setup consisted of a 33 meter by 63 meter receiver grid with 3 meter spacing in all directions with the transmitter positioned 71.5 meters from the center of the receiver grid. A total of 12 frequencies were collected from 1.1kHz to 66.2kHz. The CSAMT data is being analyzed using a 2D CSAMT RRI code (Lu, Unsworth and Booker, 1999) and a 2D MT RRI code (Smith and Booker, 1991). Of interest is examining how well the 2D codes are able to map 3D geology, the level of resolution that is obtained, and how important it is to include the 3D source in the solution. The ultimate goal is to determine the applicability of using CSAMT for mapping these types of features at the Hanford Reservation site.

  11. Formation of regularly spaced networks as a general feature of actin bundle condensation by entropic forces

    NASA Astrophysics Data System (ADS)

    Huber, Florian; Strehle, Dan; Schnauß, Jörg; Käs, Josef

    2015-04-01

    Biopolymer networks contribute mechanical integrity as well as functional organization to living cells. One of their major constituents, the protein actin, is present in a large variety of different network architectures, ranging from extensive networks to densely packed bundles. The shape of the network is directly linked to its mechanical properties and essential physiological functions. However, a profound understanding of architecture-determining mechanisms and their physical constraints remains elusive. We use experimental bottom-up systems to study the formation of confined actin networks by entropic forces. Experiments based on molecular crowding as well as counterion condensation reveal a generic tendency of homogeneous filament solutions to aggregate into regular actin bundle networks connected by aster-like centers. The network architecture is found to critically rely on network formation history. Starting from identical biochemical compositions, we observe drastic changes in network architecture as a consequence of initially biased filament orientation or mixing-induced perturbations. Our experiments suggest that the tendency to form regularly spaced bundle networks is a rather general feature of isotropic, homogeneous filament solutions subject to uniform attractive interactions. Due to the fundamental nature of the considered interactions, we expect that the investigated type of network formation further implies severe physical constraints for cytoskeleton self-organization on the more complex level of living cells.

  12. NFL Films audio, video, and film production facilities

    NASA Astrophysics Data System (ADS)

    Berger, Russ; Schrag, Richard C.; Ridings, Jason J.

    2003-04-01

    The new NFL Films 200,000 sq. ft. headquarters is home for the critically acclaimed film production that preserves the NFL's visual legacy week-to-week during the football season, and is also the technical plant that processes and archives football footage from the earliest recorded media to the current network broadcasts. No other company in the country shoots more film than NFL Films, and the inclusion of cutting-edge video and audio formats demands that their technical spaces continually integrate the latest in the ever-changing world of technology. This facility houses a staggering array of acoustically sensitive spaces where music and sound are equal partners with the visual medium. Over 90,000 sq. ft. of sound critical technical space is comprised of an array of sound stages, music scoring stages, audio control rooms, music writing rooms, recording studios, mixing theaters, video production control rooms, editing suites, and a screening theater. Every production control space in the building is designed to monitor and produce multi channel surround sound audio. An overview of the architectural and acoustical design challenges encountered for each sophisticated listening, recording, viewing, editing, and sound critical environment will be discussed.

  13. Subject-to-subject transfer for CSP based BCIs: feature space transformation and decision-level fusion.

    PubMed

    Heger, Dominic; Putze, Felix; Herff, Christian; Schultz, Tanja

    2013-01-01

    Modern Brain Computer Interfaces (BCIs) usually require a calibration session to train a machine learning system before each usage. In general, such trained systems are highly specialized to the subject's characteristic activation patterns and cannot be used for other sessions or subjects. This paper presents a feature space transformation that transforms features generated using subject-specific spatial filters into a subject-independent feature space. The transformation can be estimated from little adaptation data of the subject. Furthermore, we combine three different Common Spatial Pattern based feature extraction approaches using decision-level fusion, which enables BCI use when little calibration data is available, but also outperformed the subject-dependent reference approaches for larger amounts of training data. PMID:24111010

  14. SNR-adaptive stream weighting for audio-MES ASR.

    PubMed

    Lee, Ki-Seung

    2008-08-01

    Myoelectric signals (MESs) from the speaker's mouth region have been successfully shown to improve the noise robustness of automatic speech recognizers (ASRs), thus promising to extend their usability in implementing noise-robust ASR. In the recognition system presented herein, extracted audio and facial MES features were integrated by a decision fusion method, where the likelihood score of the audio-MES observation vector was given by a linear combination of class-conditional observation log-likelihoods of two classifiers, using appropriate weights. We developed a weighting process adaptive to SNRs. The main objective of the paper involves determining the optimal SNR classification boundaries and constructing a set of optimum stream weights for each SNR class. These two parameters were determined by a method based on a maximum mutual information criterion. Acoustic and facial MES data were collected from five subjects, using a 60-word vocabulary. Four types of acoustic noise including babble, car, aircraft, and white noise were acoustically added to clean speech signals with SNR ranging from -14 to 31 dB. The classification accuracy of the audio ASR was as low as 25.5%. Whereas, the classification accuracy of the MES ASR was 85.2%. The classification accuracy could be further improved by employing the proposed audio-MES weighting method, which was as high as 89.4% in the case of babble noise. A similar result was also found for the other types of noise. PMID:18632363

  15. Audio-visual event detection based on mining of semantic audio-visual labels

    NASA Astrophysics Data System (ADS)

    Goh, King-Shy; Miyahara, Koji; Radhakrishnan, Regunathan; Xiong, Ziyou; Divakaran, Ajay

    2003-12-01

    Removing commercials from television programs is a much sought-after feature for a personal video recorder. In this paper, we employ an unsupervised clustering scheme (CM_Detect) to detect commercials in television programs. Each program is first divided into W8-minute chunks, and we extract audio and visual features from each of these chunks. Next, we apply k-means clustering to assign each chunk with a commercial/program label. In contrast to other methods, we do not make any assumptions regarding the program content. Thus, our method is highly content-adaptive and computationally inexpensive. Through empirical studies on various content, including American news, Japanese news, and sports programs, we demonstrate that our method is able to filter out most of the commercials without falsely removing the regular program.

  16. Audio frequency analysis in mobile phones

    NASA Astrophysics Data System (ADS)

    Munguía Aguilar, Horacio

    2016-01-01

    A new experiment using mobile phones is proposed in which its audio frequency response is analyzed using the audio port for inputting external signal and getting a measurable output. This experiment shows how the limited audio bandwidth used in mobile telephony is the main cause of the poor speech quality in this service. A brief discussion is given about the relationship between voice bandwidth and voice quality.

  17. Three-Dimensional Audio Client Library

    NASA Technical Reports Server (NTRS)

    Rizzi, Stephen A.

    2005-01-01

    The Three-Dimensional Audio Client Library (3DAudio library) is a group of software routines written to facilitate development of both stand-alone (audio only) and immersive virtual-reality application programs that utilize three-dimensional audio displays. The library is intended to enable the development of three-dimensional audio client application programs by use of a code base common to multiple audio server computers. The 3DAudio library calls vendor-specific audio client libraries and currently supports the AuSIM Gold-Server and Lake Huron audio servers. 3DAudio library routines contain common functions for (1) initiation and termination of a client/audio server session, (2) configuration-file input, (3) positioning functions, (4) coordinate transformations, (5) audio transport functions, (6) rendering functions, (7) debugging functions, and (8) event-list-sequencing functions. The 3DAudio software is written in the C++ programming language and currently operates under the Linux, IRIX, and Windows operating systems.

  18. Audio-visual gender recognition

    NASA Astrophysics Data System (ADS)

    Liu, Ming; Xu, Xun; Huang, Thomas S.

    2007-11-01

    Combining different modalities for pattern recognition task is a very promising field. Basically, human always fuse information from different modalities to recognize object and perform inference, etc. Audio-Visual gender recognition is one of the most common task in human social communication. Human can identify the gender by facial appearance, by speech and also by body gait. Indeed, human gender recognition is a multi-modal data acquisition and processing procedure. However, computational multimodal gender recognition has not been extensively investigated in the literature. In this paper, speech and facial image are fused to perform a mutli-modal gender recognition for exploring the improvement of combining different modalities.

  19. Noise-Canceling Helmet Audio System

    NASA Technical Reports Server (NTRS)

    Seibert, Marc A.; Culotta, Anthony J.

    2007-01-01

    A prototype helmet audio system has been developed to improve voice communication for the wearer in a noisy environment. The system was originally intended to be used in a space suit, wherein noise generated by airflow of the spacesuit life-support system can make it difficult for remote listeners to understand the astronaut s speech and can interfere with the astronaut s attempt to issue vocal commands to a voice-controlled robot. The system could be adapted to terrestrial use in helmets of protective suits that are typically worn in noisy settings: examples include biohazard, fire, rescue, and diving suits. The system (see figure) includes an array of microphones and small loudspeakers mounted at fixed positions in a helmet, amplifiers and signal-routing circuitry, and a commercial digital signal processor (DSP). Notwithstanding the fixed positions of the microphones and loudspeakers, the system can accommodate itself to any normal motion of the wearer s head within the helmet. The system operates in conjunction with a radio transceiver. An audio signal arriving via the transceiver intended to be heard by the wearer is adjusted in volume and otherwise conditioned and sent to the loudspeakers. The wearer s speech is collected by the microphones, the outputs of which are logically combined (phased) so as to form a microphone- array directional sensitivity pattern that discriminates in favor of sounds coming from vicinity of the wearer s mouth and against sounds coming from elsewhere. In the DSP, digitized samples of the microphone outputs are processed to filter out airflow noise and to eliminate feedback from the loudspeakers to the microphones. The resulting conditioned version of the wearer s speech signal is sent to the transceiver.

  20. Recognizing the Face of Johnny, Suzy, and Me: Insensitivity to the Spacing Among Features at 4 Years of Age

    ERIC Educational Resources Information Center

    Mondloch, Catherine J.; Leis, Anishka; Maurer, Daphne

    2006-01-01

    Four-year-olds were tested for their ability to use differences in the spacing among features to recognize familiar faces. They were given a storybook depicting multiple views of 2 children. They returned to the laboratory 2 weeks later and used a "magic wand" to play a computer game that tested their ability to recognize the familiarized faces…

  1. Using virtual 3D audio in multispeech channel and multimedia environments

    NASA Astrophysics Data System (ADS)

    Orosz, Michael D.; Karplus, Walter J.; Balakrishnan, Jerry D.

    2000-08-01

    The advantages and disadvantages of using virtual 3-D audio in mission-critical, multimedia display interfaces were evaluated. The 3D audio platform seems to be an especially promising candidate for aircraft cockpits, flight control rooms, and other command and control environments in which operators must make mission-critical decisions while handling demanding and routine tasks. Virtual audio signal processing creates the illusion for a listener wearing conventional earphones that each of a multiplicity of simultaneous speech or audio channels is originating from a different, program- specified location in virtual space. To explore the possible uses of this new, readily available technology, a test bed simulating some of the conditions experienced by the chief flight test coordinator at NASA's Dryden Flight Research Center was designed and implemented. Thirty test subjects simultaneously performed routine tasks requiring constant hand-eye coordination, while monitoring four speech channels, each generating continuous speech signals, for the occurrence of pre-specified keywords. Performance measures included accuracy in identifying the keywords, accuracy in identifying the speaker of the keyword, and response time. We found substantial improvements on all of these measures when comparing virtual audio with conventional, monaural transmissions. We also explored the effect on operator performance of different spatial configurations of the audio sources in 3-D space, simulated movement (dither) in the source locations, and of providing graphical redundancy. Some of these manipulations were less effective and may even decrease performance efficiency, even though they improve some aspects of the virtual space simulation.

  2. Digital Audio Radio Field Tests

    NASA Technical Reports Server (NTRS)

    Hollansworth, James E.

    1997-01-01

    Radio history continues to be made at the NASA Lewis Research Center with the beginning of phase two of Digital Audio Radio testing conducted by the Consumer Electronic Manufacturers Association (a sector of the Electronic Industries Association and the National Radio Systems Committee) and cosponsored by the Electronic Industries Association and the National Association of Broadcasters. The bulk of the field testing of the four systems should be complete by the end of October 1996, with results available soon thereafter. Lewis hosted phase one of the testing process, which included laboratory testing of seven proposed digital audio radio systems and modes (see the following table). Two of the proposed systems operate in two modes, thus making a total of nine systems for testing. These nine systems are divided into the following types of transmission: in-band on channel (IBOC), in-band adjacent channel (IBAC), and new bands - the L-band (1452 to 1492 MHz) and the S-band (2310 to 2360 MHz).

  3. Instructional Audio Communications at a Glance.

    ERIC Educational Resources Information Center

    Alaska Univ., Anchorage.

    Audio conferencing, or teleconferencing, is fast becoming an important method of delivering education to students at a distance. Audio conferencing equipment consists of a convener (speaker) and push-to-talk microphone. Use of the system costs less than using a number of separate telephone lines and is more comfortable for users. Advantages of…

  4. Dual Audio TV Instruction: A Broadcast Experiment.

    ERIC Educational Resources Information Center

    Borton, Terry; And Others

    An experiment assessed the potential effectiveness of "dual audio television instruction" (DATI) as a mass education medium. The DATI consisted of a radio program heard by children while they watched television shows. The audio instructor did not talk when the television characters spoke, but used the "quiet" times to help with reading, define…

  5. Digital Audio: A Sound Design Element.

    ERIC Educational Resources Information Center

    Barron, Ann; Varnadoe, Susan

    1992-01-01

    Discussion of incorporating audio into videodiscs for multimedia educational applications highlights a project developed for the Navy that used digital audio in an interactive video delivery system (IVDS) for training sonar operators. Storage constraints with videodiscs are explained, design requirements for the IVDS are described, and production…

  6. Audio-Visual Aids in Universities

    ERIC Educational Resources Information Center

    Douglas, Jackie

    1970-01-01

    A report on the proceedings and ideas expressed at a one day seminar on "Audio-Visual Equipment--Its Uses and Applications for Teaching and Research in Universities." The seminar was organized by England's National Committee for Audio-Visual Aids in Education in conjunction with the British Universities Film Council. (LS)

  7. The Concept of Audio-Tutorial Teaching.

    ERIC Educational Resources Information Center

    Husband, D. D.; Postlethwait, S. N.

    Audio-tutorial teaching, originated at Purdue University, employs an audio tape as a vehicle for guiding the student through a series of learning experiences. The student may be directed to use his textbook for an explanation of a diagram, to refer to his laboratory manual, or to observe the specimens and experimental materials prepared for his…

  8. Digital Audio Sampling for Film and Video.

    ERIC Educational Resources Information Center

    Stanton, Michael J.

    Digital audio sampling is explained, and some of its implications in digital sound applications are discussed. Digital sound equipment is rapidly replacing analog recording devices as the state-of-the-art in audio technology. The philosophy of digital recording involves doing away with the continuously variable analog waveforms and turning the…

  9. Audio-Tutorial Instruction in Medicine.

    ERIC Educational Resources Information Center

    Boyle, Gloria J.; Herrick, Merlyn C.

    This progress report concerns an audio-tutorial approach used at the University of Missouri-Columbia School of Medicine. Instructional techniques such as slide-tape presentations, compressed speech audio tapes, computer-assisted instruction (CAI), motion pictures, television, microfiche, and graphic and printed materials have been implemented,…

  10. Internet Audio Products (3/3)

    ERIC Educational Resources Information Center

    Schwartz, Linda; de Schutter, Adrienne; Fahrni, Patricia; Rudolph, Jim

    2004-01-01

    Two contrasting additions to the online audio market are reviewed: "iVocalize", a browser-based audio-conferencing software, and "Skype", a PC-to-PC Internet telephone tool. These products are selected for review on the basis of their success in gaining rapid popular attention and usage during 2003-04. The "iVocalize" review emphasizes the…

  11. Improvements of ModalMax High-Fidelity Piezoelectric Audio Device

    NASA Technical Reports Server (NTRS)

    Woodard, Stanley E.

    2005-01-01

    ModalMax audio speakers have been enhanced by innovative means of tailoring the vibration response of thin piezoelectric plates to produce a high-fidelity audio response. The ModalMax audio speakers are 1 mm in thickness. The device completely supplants the need to have a separate driver and speaker cone. ModalMax speakers can perform the same applications of cone speakers, but unlike cone speakers, ModalMax speakers can function in harsh environments such as high humidity or extreme wetness. New design features allow the speakers to be completely submersed in salt water, making them well suited for maritime applications. The sound produced from the ModalMax audio speakers has sound spatial resolution that is readily discernable for headset users.

  12. Robot Command Interface Using an Audio-Visual Speech Recognition System

    NASA Astrophysics Data System (ADS)

    Ceballos, Alexánder; Gómez, Juan; Prieto, Flavio; Redarce, Tanneguy

    In recent years audio-visual speech recognition has emerged as an active field of research thanks to advances in pattern recognition, signal processing and machine vision. Its ultimate goal is to allow human-computer communication using voice, taking into account the visual information contained in the audio-visual speech signal. This document presents a command's automatic recognition system using audio-visual information. The system is expected to control the laparoscopic robot da Vinci. The audio signal is treated using the Mel Frequency Cepstral Coefficients parametrization method. Besides, features based on the points that define the mouth's outer contour according to the MPEG-4 standard are used in order to extract the visual speech information.

  13. Linking molecular feature space and disease terms for the immunosuppressive drug rapamycin.

    PubMed

    Bernthaler, Andreas; Mönks, Konrad; Mühlberger, Irmgard; Mayer, Bernd; Perco, Paul; Oberbauer, Rainer

    2011-10-01

    Next to development of novel drugs also drug repositioning appears promising for tackling unmet clinical needs. Here Omics provided the ground for novel analysis strategies for linking drug and disease by integrating profiles on the molecular as well as the clinical data level. We developed a workflow for linking drugs and diseases for identifying repositioning options, and exemplify the procedure for the immunosuppressive drug rapamycin. Our strategy rests on delineating a drug-specific molecular profile by combining Omics data reflecting the drug's impact on the cellular status as well as drug-associated molecular features extracted from the scientific literature. For rapamycin the respective profile held 905 unique molecular features reflecting defined molecular processes as identified by molecular pathway and process enrichment analysis. Literature mining identified 419 diseases significantly associated with this rapamycin molecular feature list, and transforming the significance of gene-disease associations into a continuous score allowed us to compute ROC and precision-recall for comparing this disease list with diseases already undergoing clinical trials utilizing rapamycin. The AUC of this assignment was computed as 0.84, indicating excellent recovery of relevant disease terms solely based on the drug molecular feature profile. We verified relevant indications by comparing molecular feature sets characteristic for the identified diseases to the drug molecular feature profile, demonstrating highly significant overlaps. The presented workflow allowed positive identification of diseases associated with rapamycin utilizing the drug-specific molecular feature profile, and may be well applicable to other drugs of interest. PMID:21789336

  14. Authenticity examination of compressed audio recordings using detection of multiple compression and encoders' identification.

    PubMed

    Korycki, Rafal

    2014-05-01

    Since the appearance of digital audio recordings, audio authentication has been becoming increasingly difficult. The currently available technologies and free editing software allow a forger to cut or paste any single word without audible artifacts. Nowadays, the only method referring to digital audio files commonly approved by forensic experts is the ENF criterion. It consists in fluctuation analysis of the mains frequency induced in electronic circuits of recording devices. Therefore, its effectiveness is strictly dependent on the presence of mains signal in the recording, which is a rare occurrence. Recently, much attention has been paid to authenticity analysis of compressed multimedia files and several solutions were proposed for detection of double compression in both digital video and digital audio. This paper addresses the problem of tampering detection in compressed audio files and discusses new methods that can be used for authenticity analysis of digital recordings. Presented approaches consist in evaluation of statistical features extracted from the MDCT coefficients as well as other parameters that may be obtained from compressed audio files. Calculated feature vectors are used for training selected machine learning algorithms. The detection of multiple compression covers up tampering activities as well as identification of traces of montage in digital audio recordings. To enhance the methods' robustness an encoder identification algorithm was developed and applied based on analysis of inherent parameters of compression. The effectiveness of tampering detection algorithms is tested on a predefined large music database consisting of nearly one million of compressed audio files. The influence of compression algorithms' parameters on the classification performance is discussed, based on the results of the current study. PMID:24637036

  15. PSOFuzzySVM-TMH: identification of transmembrane helix segments using ensemble feature space by incorporated fuzzy support vector machine.

    PubMed

    Hayat, Maqsood; Tahir, Muhammad

    2015-08-01

    Membrane protein is a central component of the cell that manages intra and extracellular processes. Membrane proteins execute a diversity of functions that are vital for the survival of organisms. The topology of transmembrane proteins describes the number of transmembrane (TM) helix segments and its orientation. However, owing to the lack of its recognized structures, the identification of TM helix and its topology through experimental methods is laborious with low throughput. In order to identify TM helix segments reliably, accurately, and effectively from topogenic sequences, we propose the PSOFuzzySVM-TMH model. In this model, evolutionary based information position specific scoring matrix and discrete based information 6-letter exchange group are used to formulate transmembrane protein sequences. The noisy and extraneous attributes are eradicated using an optimization selection technique, particle swarm optimization, from both feature spaces. Finally, the selected feature spaces are combined in order to form ensemble feature space. Fuzzy-support vector Machine is utilized as a classification algorithm. Two benchmark datasets, including low and high resolution datasets, are used. At various levels, the performance of the PSOFuzzySVM-TMH model is assessed through 10-fold cross validation test. The empirical results reveal that the proposed framework PSOFuzzySVM-TMH outperforms in terms of classification performance in the examined datasets. It is ascertained that the proposed model might be a useful and high throughput tool for academia and research community for further structure and functional studies on transmembrane proteins. PMID:26054033

  16. Interference from audio distracters during speechreading

    NASA Astrophysics Data System (ADS)

    Brungart, Douglas S.; Simpson, Brian D.

    2005-12-01

    Although many audio-visual speech experiments have focused on situations where the presence of an incongruent visual speech signal influences the perceived utterance heard by an observer, there are also documented examples of a related effect in which the presence of an incongruent audio speech signal influences the perceived utterance seen by an observer. This study examined the effects that different distracting audio signals had on performance in a color and number keyword speechreading task. When the distracting sound was noise, time-reversed speech, or continuous speech, it had no effect on speechreading. However, when the distracting audio signal consisted of speech that started at the same time as the visual stimulus, speechreading performance was substantially degraded. This degradation did not depend on the semantic similarity between the target and masker speech, but it was substantially reduced when the onset of the audio speech was shifted relative to that of the visual stimulus. Overall, these results suggest that visual speech perception is impaired by the presence of a simultaneous mismatched audio speech signal, but that other types of audio distracters have little effect on speechreading performance.

  17. High-Fidelity Piezoelectric Audio Device

    NASA Technical Reports Server (NTRS)

    Woodward, Stanley E.; Fox, Robert L.; Bryant, Robert G.

    2003-01-01

    ModalMax is a very innovative means of harnessing the vibration of a piezoelectric actuator to produce an energy efficient low-profile device with high-bandwidth high-fidelity audio response. The piezoelectric audio device outperforms many commercially available speakers made using speaker cones. The piezoelectric device weighs substantially less (4 g) than the speaker cones which use magnets (10 g). ModalMax devices have extreme fabrication simplicity. The entire audio device is fabricated by lamination. The simplicity of the design lends itself to lower cost. The piezoelectric audio device can be used without its acoustic chambers and thereby resulting in a very low thickness of 0.023 in. (0.58 mm). The piezoelectric audio device can be completely encapsulated, which makes it very attractive for use in wet environments. Encapsulation does not significantly alter the audio response. Its small size (see Figure 1) is applicable to many consumer electronic products, such as pagers, portable radios, headphones, laptop computers, computer monitors, toys, and electronic games. The audio device can also be used in automobile or aircraft sound systems.

  18. Neuromagnetic Evidence for a Featural Distinction of English Consonants: Sensor- and Source-Space Data

    ERIC Educational Resources Information Center

    Scharinger, Mathias; Merickel, Jennifer; Riley, Joshua; Idsardi, William J.

    2011-01-01

    Speech sounds can be classified on the basis of their underlying articulators or on the basis of the acoustic characteristics resulting from particular articulatory positions. Research in speech perception suggests that distinctive features are based on both articulatory and acoustic information. In recent years, neuroelectric and neuromagnetic…

  19. Audio recording and reproduction in CARROUSO: Getting closer to perfection?

    NASA Astrophysics Data System (ADS)

    Teutsch, Heinz; Spors, Sascha; Buchner, Herbert; Rabenstein, Rudolf; Kellermann, Walter

    2002-05-01

    State-of-the-art systems for spatial audio reproduction utilize two to six discrete playback channels. A problem inherent to these systems is the relatively small area where the listener is able to experience a true 3-D sound sensation. This so-called ``sweet spot'' can be significantly enlarged by using loudspeaker arrays in combination with wave field synthesis (WFS) technology, initially developed at Delft University. By following this approach, actual sonic spaces can be reproduced in their entirety and not only discrete multichannel representations thereof. While loudspeaker arrays can be used to reproduce sound fields, microphone arrays can be used for sound field capture and analysis. Having high-quality audio reproduction in mind, microphone array designs are presented that need to fulfill stricter requirements than what has been traditionally considered for microphone array applications. Information on acoustic source position is essential for WFS-based rendering techniques. As will be shown, joint audio-video object tracking proves to be efficient for this task. Moreover, full-duplex applications based on WFS technology, like high-quality teleconferencing or remote music teaching, call for sophisticated multichannel acoustic echo cancellation algorithms. The European project ``CARROUSO'' aims at developing, integrating, and building a real-time system that embraces all previously described technologies in an MPEG-4 context.

  20. Perceived Danger in Urban Public Space. The Impacts of Physical Features and Personal Factors

    ERIC Educational Resources Information Center

    Blobaum, Anke; Hunecke, Marcel

    2005-01-01

    What are the most relevant factors influencing perceived danger in urban public space? To answer this question, a field experiment of students(N = 122) was carried out on a German university campus within which perceived danger was analyzed under systematic variation of lighting, prospect, and opportunities of escape. Two standardized…

  1. Virtual environment display for a 3D audio room simulation

    NASA Astrophysics Data System (ADS)

    Chapin, William L.; Foster, Scott

    1992-06-01

    Recent developments in virtual 3D audio and synthetic aural environments have produced a complex acoustical room simulation. The acoustical simulation models a room with walls, ceiling, and floor of selected sound reflecting/absorbing characteristics and unlimited independent localizable sound sources. This non-visual acoustic simulation, implemented with 4 audio ConvolvotronsTM by Crystal River Engineering and coupled to the listener with a Poihemus IsotrakTM, tracking the listener's head position and orientation, and stereo headphones returning binaural sound, is quite compelling to most listeners with eyes closed. This immersive effect should be reinforced when properly integrated into a full, multi-sensory virtual environment presentation. This paper discusses the design of an interactive, visual virtual environment, complementing the acoustic model and specified to: 1) allow the listener to freely move about the space, a room of manipulable size, shape, and audio character, while interactively relocating the sound sources; 2) reinforce the listener's feeling of telepresence into the acoustical environment with visual and proprioceptive sensations; 3) enhance the audio with the graphic and interactive components, rather than overwhelm or reduce it; and 4) serve as a research testbed and technology transfer demonstration. The hardware/software design of two demonstration systems, one installed and one portable, are discussed through the development of four iterative configurations. The installed system implements a head-coupled, wide-angle, stereo-optic tracker/viewer and multi-computer simulation control. The portable demonstration system implements a head-mounted wide-angle, stereo-optic display, separate head and pointer electro-magnetic position trackers, a heterogeneous parallel graphics processing system, and object oriented C++ program code.

  2. Nonlinear color space and spatiotemporal MRF for hierarchical segmentation of face features in video.

    PubMed

    Liévin, Marc; Luthon, Franck

    2004-01-01

    This paper deals with the low-level joint processing of color and motion for robust face analysis within a feature-based approach. To gain robustness and contrast under unsupervised viewing conditions, a nonlinear color transform relevant for hue segmentation is derived from a logarithmic model. A hierarchical segmentation scheme is based on Markov random field modeling, that combines hue and motion detection within a spatiotemporal neighborhood. Relevant face regions are segmented without parameter tuning. The accuracy of the label fields enables not only face detection and tracking but also geometrical measurements on facial feature edges, such as lips or eyes. Results are shown both on typical test sequences and on various sequences acquired from micro- or mobile cameras. The efficiency of the method makes it suitable for real-time applications aiming at audiovisual communication in unsupervised environments. PMID:15376958

  3. A Study of Audio Tape: Part II

    ERIC Educational Resources Information Center

    Reen, Noel K.

    1975-01-01

    To evaluate reel audio tape, tests were performed to identify: signal-to-noise ratio, total harmonic distortion, dynamic response, frequency response, biased and virgin tape noise, dropout susceptibility and oxide coating uniformity. (SCC)

  4. Audio-visual affective expression recognition

    NASA Astrophysics Data System (ADS)

    Huang, Thomas S.; Zeng, Zhihong

    2007-11-01

    Automatic affective expression recognition has attracted more and more attention of researchers from different disciplines, which will significantly contribute to a new paradigm for human computer interaction (affect-sensitive interfaces, socially intelligent environments) and advance the research in the affect-related fields including psychology, psychiatry, and education. Multimodal information integration is a process that enables human to assess affective states robustly and flexibly. In order to understand the richness and subtleness of human emotion behavior, the computer should be able to integrate information from multiple sensors. We introduce in this paper our efforts toward machine understanding of audio-visual affective behavior, based on both deliberate and spontaneous displays. Some promising methods are presented to integrate information from both audio and visual modalities. Our experiments show the advantage of audio-visual fusion in affective expression recognition over audio-only or visual-only approaches.

  5. A study of some features of ac and dc electric power systems for a space station

    NASA Technical Reports Server (NTRS)

    Hanania, J. I.

    1983-01-01

    This study analyzes certain selected topics in rival dc and high frequency ac electric power systems for a Space Station. The interaction between the Space Station and the plasma environment is analyzed, leading to a limit on the voltage for the solar array and a potential problem with resonance coupling at high frequencies. Certain problems are pointed out in the concept of a rotary transformer, and further development work is indicated in connection with dc circuit switching, special design of a transmission conductor for the ac system, and electric motors. The question of electric shock hazards, particularly at high frequency, is also explored. and a problem with reduced skin resistance and therefore increased hazard with high frequency ac is pointed out. The study concludes with a comparison of the main advantages and disadvantages of the two rival systems, and it is suggested that the choice between the two should be made after further studies and development work are completed.

  6. Ophthalmic Features of Outpatient Children Diagnosed with Intracranial Space-Occupying Lesions by Ophthalmologists

    PubMed Central

    Alswaina, Nayef; Elkhamary, Sahar M.; Shammari, Mansour A.; Khan, Arif O.

    2015-01-01

    Introduction: Brain tumors in children often involve the visual system, but most retrospective series are by neurologists or oncologists. In this study we highlight the ophthalmic findings of outpatient children with visual complaints and/or strabismus who, based on ophthalmic examination, were suspected to and confirmed to harbor intracranial space-occupying lesions by magnetic resonance imaging (MRI). Materials and Methods: Retrospective case series of children (less than 18 years) who for visual complaints and/or strabismus underwent cranial MRI at a referral eye hospital (2005–2012), which revealed intracranial space-occupying lesions. Exclusion criteria were known preexisting orbital or ocular trauma, ocular tumor, or neurological disease. Results: For 26 patients (3 months-17 years; mean 7 years; median 9 years; and 14 boys), the most common clinical presentation was decreased vision with disc pallor (10) or swelling (three). Other presentations were strabismus with disc pallor or swelling (four; two of which were left sixth nerve palsies), acquired esotropia with diplopia (three; one bilateral and two left sixth nerve palsies), acquired exotropia (four; two of which were bilateral third nerve palsies, one of which was left partial third nerve palsy, and one of which was associated with headache), nystagmus (one), and disc swelling with headache (one). Most lesions were in the sellar/suprasellar space (10), posterior fossa (six), or optic nerve/chasm (four). Conclusions: The majority of outpatient children diagnosed by ophthalmologists with intracranial space-occupying lesions presented with disc swelling or pallor in the context of decreased vision or strabismus. Two strabismus profiles that did not include disc swelling or pallor were acquired sixth nerve palsy and acquired exotropia (with ptosis (third nerve palsy), nystagmus, or headache). PMID:26180471

  7. Micromorphometric features of pore space in the plow horizons of loamy soils

    NASA Astrophysics Data System (ADS)

    Skvortsova, E. B.; Sanzharova, S. I.

    2007-04-01

    A computer-based analysis of thin sections has been applied to study pore space in the plow horizons of loamy soils in European Russia and Ukraine. Differences in the morphology of soil macro-and mesopores are shown. It is argued that agrogenic impacts result in the convergence of the shape and orientation of macropores in plowed loamy soils of the forest, forest-steppe, and steppe zones. At the same time, this convergence is not observed for the soil mesopores.

  8. The features of radiation dose variations onboard ISS and Mir space station: comparative study.

    PubMed

    Tverskaya, L V; Panasyuk, M I; Reizman, S Ya; Sosnovets, E N; Teltsov, M V; Tsetlin, V V

    2004-01-01

    The dynamics of the ISS-measured radiation dose variations since August 2000 is studied. Use is made of the data obtained with the R-16 instrument, which consists of two ionization chambers behind different shielding thicknesses. The doses recorded during solar energetic particle (SEP) events are compared with the data obtained also by R-16 on Mir space station. The SEP events in the solar maximum of the current cycle make a much smaller contribution to the radiation dose compared with the October 1989 event recorded on Mir space station. In the latter event, the proton intensity was peaking during a strong magnetic storm. The storm-time effect of solar proton geomagnetic cutoff decreases on dose variations is estimated. The dose variations on Mir space stations due to formation of a new radiation belt of high-energy protons and electrons during a sudden commencement of March 24, 1991 storm are also studied. It was for the first time throughout the ISS and Mir dose measurement period that the counting rates recorded by both R-16 channels on ISS in 2001-2002 were nearly the same during some time intervals. This effect may arise from the decreases of relativistic electron fluxes in the outer radiation belt. PMID:15881786

  9. Huffman coding in advanced audio coding standard

    NASA Astrophysics Data System (ADS)

    Brzuchalski, Grzegorz

    2012-05-01

    This article presents several hardware architectures of Advanced Audio Coding (AAC) Huffman noiseless encoder, its optimisations and working implementation. Much attention has been paid to optimise the demand of hardware resources especially memory size. The aim of design was to get as short binary stream as possible in this standard. The Huffman encoder with whole audio-video system has been implemented in FPGA devices.

  10. Self-synchronization for spread spectrum audio watermarks after time scale modification

    NASA Astrophysics Data System (ADS)

    Nadeau, Andrew; Sharma, Gaurav

    2014-02-01

    De-synchronizing operations such as insertion, deletion, and warping pose significant challenges for watermarking. Because these operations are not typical for classical communications, watermarking techniques such as spread spectrum can perform poorly. Conversely, specialized synchronization solutions can be challenging to analyze/ optimize. This paper addresses desynchronization for blind spread spectrum watermarks, detected without reference to any unmodified signal, using the robustness properties of short blocks. Synchronization relies on dynamic time warping to search over block alignments to find a sequence with maximum correlation to the watermark. This differs from synchronization schemes that must first locate invariant features of the original signal, or estimate and reverse desynchronization before detection. Without these extra synchronization steps, analysis for the proposed scheme builds on classical SS concepts and allows characterizes the relationship between the size of search space (number of detection alignment tests) and intrinsic robustness (continuous search space region covered by each individual detection test). The critical metrics that determine the search space, robustness, and performance are: time-frequency resolution of the watermarking transform, and blocklength resolution of the alignment. Simultaneous robustness to (a) MP3 compression, (b) insertion/deletion, and (c) time-scale modification is also demonstrated for a practical audio watermarking scheme developed in the proposed framework.

  11. Competing descriptions of diffusion profiles with two features: Surface space-charge layer versus fast grain-boundary diffusion

    NASA Astrophysics Data System (ADS)

    Schraknepper, H.; De Souza, R. A.

    2016-02-01

    Two different physical processes, (i) fast grain-boundary diffusion (FGBD) of oxygen and (ii) hindered oxygen diffusion in a surface space-charge layer, yield oxygen isotope diffusion profiles in a similar form. Two features are observed, with the short, sharp profile close to the surface being followed by a longer, shallower profile. In this study, we develop a procedure for deciding which of the two descriptions applies to experimentally measured profiles. Specifically, we solve Fick's second law, using finite-element simulations, to obtain oxygen isotope diffusion profiles for the two cases. Each set of profiles is then analysed in terms of the competing description. In this manner, we derive falsifiable conditions that allow physical processes to be assigned unambiguously to the two features of such isotope profiles. Applying these conditions to experimental profiles for SrTiO3 single crystals published in the literature, we find that FGBD is an invalid model for describing the diffusion processes.

  12. Miniaturized, 9-12 micron heterodyne spectrometer with space qualifiable design features

    NASA Technical Reports Server (NTRS)

    Glenar, D. A.; Mumma, M. J.; Kostiuk, T.; Huffman, H.; Degnan, J.

    1990-01-01

    A demonstration-prototype CO2-laser heterodyne spectrometer operating at 9-12 microns and suitable for long-term space missions is described and illustrated with extensive diagrams, drawings, photographs, and graphs of test performance data. The spectrometer has total volume 0.63 cu m, mass 30 kg, and power requirement 60-70 W, compatible with miniature-class Space Shuttle experiment payload specifications. It comprises three modules: (1) an optical front end with reflecting optics, a 2-GHz BW HgCdTe photomixer, and a 0-2-GHz 40-dB RF preamplifier; (2) a local oscillator with an RF-excited waveguide CO2 laser, a 75-percent-efficiency RF amplifier, a stepper-driven grating mode selector, and an etalon stabilized for over 30,000 h of use; and (3) an RF-filter-bank spectral-line receiver with a 25-MHz RF channel, 1.6-GHz IF spectral coverage, onboard instrument control, a serial link to the host computer, and highly integrated design.

  13. A Visual Analytics Approach Using the Exploration of Multidimensional Feature Spaces for Content-Based Medical Image Retrieval.

    PubMed

    Kumar, Ashnil; Nette, Falk; Klein, Karsten; Fulham, Michael; Kim, Jinman

    2015-09-01

    Content-based image retrieval (CBIR) is a search technique based on the similarity of visual features and has demonstrated potential benefits for medical diagnosis, education, and research. However, clinical adoption of CBIR is partially hindered by the difference between the computed image similarity and the user's search intent, the semantic gap, with the end result that relevant images with outlier features may not be retrieved. Furthermore, most CBIR algorithms do not provide intuitive explanations as to why the retrieved images were considered similar to the query (e.g., which subset of features were similar), hence, it is difficult for users to verify if relevant images, with a small subset of outlier features, were missed. Users, therefore, resort to examining irrelevant images and there are limited opportunities to discover these "missed" images. In this paper, we propose a new approach to medical CBIR by enabling a guided visual exploration of the search space through a tool, called visual analytics for medical image retrieval (VAMIR). The visual analytics approach facilitates interactive exploration of the entire dataset using the query image as a point-of-reference. We conducted a user study and several case studies to demonstrate the capabilities of VAMIR in the retrieval of computed tomography images and multimodality positron emission tomography and computed tomography images. PMID:25296409

  14. Special Features of the Structure of Secular Resonances in the Dynamics of Near-Earth Space Objects

    NASA Astrophysics Data System (ADS)

    Bordovitsyna, T. V.; Tomilova, I. V.

    2016-07-01

    The special features of the structure of secular resonances in the near-earth orbital space bounded by the following range of orbital parameters: semimajor axis from 8000 to 55 000 km, inclination from 0 to 90°, and eccentricity equal to 0.01, 0.6, and 0.8 are analyzed. The influence of stable and unstable secular resonances on the long-term orbital evolution of near-earth space objects is also considered. It is demonstrated that the joint effect of the stable secular resonances of different spectral classes does not violate the regularity of motion. The chaoticity arises when stable secular resonances of one spectral class are imposed.

  15. Our Understanding of Space Weather features responsible for geostationary satellite anamolies (P39)

    NASA Astrophysics Data System (ADS)

    Rajaram, G.; et al.

    2006-11-01

    girija60@yahoo.com The topic of space weather effects on operational anomalies on spacecraft is one of considerable research investigation, with both pure and applied aspects. This is because of the very high costs involved in fabricating and operating spacecraft, and in insuring them against the harmful effects of space weather. This is more true for geostationary satellites than of low-orbiting spacecraft, as the former operate in the high-risk environment of the Earth’s outer radiation belts, with its large vagaries in spatial and time variations of high- energy electron and proton distributions (BAKER 2003). Without doubt, plasma and magnetic field emissions from active regions on the Sun are the root cause for spacecraft anomalies. Our study for 2005 shows that over 95% of anomalies can be related to some definite activity on the Sun, ranging from high-speed solar wind streams with their 27-day recurrence patterns/coronal holes/coronal mass ejections preceded by X or M type of flares/and magnetic cloud events. The most energetic solar activity events are generally accompanied by a large rise in solar MeV proton densities at geo-stationary orbit (WILKINSON 1994), and they account for definite anomalies classified as SEU (Single Event Upsets which most often are reversible through resetting commands). Any particles in the low energy ranges (eV and keV, and these could be of magnetospheric or ionospheric origin), are believed to cause external charging effects in exposed parts of the spacecraft such as solar power arrays and power cables. These mainly result in power losses which are debilitating over a period of time. The most dangerous and often irrecoverable damage is due to electronics in the 1-5 MeV range which cause deep dielectric discharge of arc type in semi-conductors comprising spacecraft instruments. Following major solar activity, the populations of these rise to more than (5x103) particles/cm2.ster.sec, with large spatial and time variations (LOVE

  16. The research on image encryption method based on parasitic audio watermark

    NASA Astrophysics Data System (ADS)

    Gao, Pei-pei; Zhu, Yao-ting; Zhang, Shi-tao

    2010-11-01

    In order to improve image encryption strength, an image encryption method based on parasitic audio watermark was proposed in this paper, which relies on double messages such as image domain and speech domain to do image encryption protection. The method utilizes unique Chinese phonetics synthesis algorithm to complete audio synthesis with embedded text, then separate this sentence information into prosodic phrase, obtains complete element set of initial consonant and compound vowel that reflects audio feature of statement. By sampling and scrambling the initial consonant and compound vowel element, synthesizing them with image watermark, and embedding the compound into the image to be encrypted in frequency domain, the processed image contains image watermark information and parasitizes audio feature information. After watermark extraction, using the same phonetics synthesis algorithm the audio information is synthesized and compared with the original. Experiments show that any decryption method in image domain or speech domain could not break encryption protection and image gains higher encryption strength and security level by double encryption.

  17. Could Audio-Described Films Benefit from Audio Introductions? An Audience Response Study

    ERIC Educational Resources Information Center

    Romero-Fresco, Pablo; Fryer, Louise

    2013-01-01

    Introduction: Time constraints limit the quantity and type of information conveyed in audio description (AD) for films, in particular the cinematic aspects. Inspired by introductory notes for theatre AD, this study developed audio introductions (AIs) for "Slumdog Millionaire" and "Man on Wire." Each AI comprised 10 minutes of…

  18. Spatial domain entertainment audio decompression/compression

    NASA Astrophysics Data System (ADS)

    Chan, Y. K.; Tam, Ka Him K.

    2014-02-01

    The ARM7 NEON processor with 128bit SIMD hardware accelerator requires a peak performance of 13.99 Mega Cycles per Second for MP3 stereo entertainment quality decoding. For similar compression bit rate, OGG and AAC is preferred over MP3. The Patent Cooperation Treaty Application dated 28/August/2012 describes an audio decompression scheme producing a sequence of interleaving "min to Max" and "Max to min" rising and falling segments. The number of interior audio samples bound by "min to Max" or "Max to min" can be {0|1|…|N} audio samples. The magnitudes of samples, including the bounding min and Max, are distributed as normalized constants within the 0 and 1 of the bounding magnitudes. The decompressed audio is then a "sequence of static segments" on a frame by frame basis. Some of these frames needed to be post processed to elevate high frequency. The post processing is compression efficiency neutral and the additional decoding complexity is only a small fraction of the overall decoding complexity without the need of extra hardware. Compression efficiency can be speculated as very high as source audio had been decimated and converted to a set of data with only "segment length and corresponding segment magnitude" attributes. The PCT describes how these two attributes are efficiently coded by the PCT innovative coding scheme. The PCT decoding efficiency is obviously very high and decoding latency is basically zero. Both hardware requirement and run time is at least an order of magnitude better than MP3 variants. The side benefit is ultra low power consumption on mobile device. The acid test on how such a simplistic waveform representation can indeed reproduce authentic decompressed quality is benchmarked versus OGG(aoTuv Beta 6.03) by three pair of stereo audio frames and one broadcast like voice audio frame with each frame consisting 2,028 samples at 44,100KHz sampling frequency.

  19. Multimodal audio guide for museums and exhibitions

    NASA Astrophysics Data System (ADS)

    Gebbensleben, Sandra; Dittmann, Jana; Vielhauer, Claus

    2006-02-01

    In our paper we introduce a new Audio Guide concept for exploring buildings, realms and exhibitions. Actual proposed solutions work in most cases with pre-defined devices, which users have to buy or borrow. These systems often go along with complex technical installations and require a great degree of user training for device handling. Furthermore, the activation of audio commentary related to the exhibition objects is typically based on additional components like infrared, radio frequency or GPS technology. Beside the necessity of installation of specific devices for user location, these approaches often only support automatic activation with no or limited user interaction. Therefore, elaboration of alternative concepts appears worthwhile. Motivated by these aspects, we introduce a new concept based on usage of the visitor's own mobile smart phone. The advantages in our approach are twofold: firstly the Audio Guide can be used in various places without any purchase and extensive installation of additional components in or around the exhibition object. Secondly, the visitors can experience the exhibition on individual tours only by uploading the Audio Guide at a single point of entry, the Audio Guide Service Counter, and keeping it on her or his personal device. Furthermore, since the user usually is quite familiar with the interface of her or his phone and can thus interact with the application device easily. Our technical concept makes use of two general ideas for location detection and activation. Firstly, we suggest an enhanced interactive number based activation by exploiting the visual capabilities of modern smart phones and secondly we outline an active digital audio watermarking approach, where information about objects are transmitted via an analog audio channel.

  20. High-energy electromagnetic cascades in extragalactic space: Physics and features

    NASA Astrophysics Data System (ADS)

    Berezinsky, V.; Kalashev, O.

    2016-07-01

    Using the analytic modeling of the electromagnetic cascades compared with more precise numerical simulations, we describe the physical properties of electromagnetic cascades developing in the universe on cosmic microwave background and extragalactic background light radiations. A cascade is initiated by very-high-energy photon or electron, and the remnant photons at large distance have two-component energy spectrum, ∝E-2 (∝E-1.9 in numerical simulations) produced at the cascade multiplication stage and ∝E-3 /2 from Inverse Compton electron cooling at low energies. The most noticeable property of the cascade spectrum in analytic modeling is "strong universality," which includes the standard energy spectrum and the energy density of the cascade ωcas as its only numerical parameter. Using numerical simulations of the cascade spectrum and comparing it with recent Fermi LAT spectrum, we obtained the upper limit on ωcas stronger than in previous works. The new feature of the analysis is the "Emax rule." We investigate the dependence of ωcas on the distribution of sources, distinguishing two cases of universality: the strong and weak ones.

  1. Features of space-charge-limited emission in foil-less diodes

    SciTech Connect

    Wu, Ping; Yuan, Keliang; Liu, Guozhi; Sun, Jun

    2014-12-15

    Space-charge-limited (SCL) current can always be obtained from the blade surface of annular cathodes in foil-less diodes which are widely used in O-type relativistic high power microwave generators. However, there is little theoretical analysis regarding it due to the mathematical complexity, and almost all formulas about the SCL current in foil-less diodes are based on numerical simulation results. This paper performs an initial trial in calculation of the SCL current from annular cathodes theoretically under the ultra-relativistic assumption and the condition of infinitely large guiding magnetic field. The numerical calculation based on the theoretical research is coherent with the particle-in-cell (PIC) simulation result to some extent under a diode voltage of 850 kV. Despite that the theoretical research gives a much larger current than the PIC simulation (41.3 kA for the former and 9.7 kA for the latter), which is induced by the ultra-relativistic assumption in the theoretical research, they both show the basic characteristic of emission from annular cathodes in foil-less diodes, i.e., the emission enhancement at the cathode blade edges, especially at the outer edge. This characteristic is confirmed to some extent in our experimental research of cathode plasma photographing under the same diode voltage and a guiding magnetic field of 4 T.

  2. Estimating Species Distributions Across Space Through Time and with Features of the Environment

    SciTech Connect

    Kelling, S.; Fink, D.; Hochachka, W.; Rosenberg, K.; Cook, R.; Damoulas, C.; Silva, C.; Michener, W.

    2013-01-01

    Complete guidance for mastering the tools and techniques of the digital revolution With the digital revolution opening up tremendous opportunities in many fields, there is a growing need for skilled professionals who can develop data-intensive systems and extract information and knowledge from them. This book frames for the first time a new systematic approach for tackling the challenges of data-intensive computing, providing decision makers and technical experts alike with practical tools for dealing with our exploding data collections. Emphasizing data-intensive thinking and interdisciplinary collaboration, The Data Bonanza: Improving Knowledge Discovery in Science, Engineering, and Business examines the essential components of knowledge discovery, surveys many of the current research efforts worldwide, and points to new areas for innovation. Complete with a wealth of examples and DISPEL-based methods demonstrating how to gain more from data in real-world systems, the book: Outlines the concepts and rationale for implementing data-intensive computing in organizations Covers from the ground up problem-solving strategies for data analysis in a data-rich world Introduces techniques for data-intensive engineering using the Data-Intensive Systems Process Engineering Language DISPEL Features in-depth case studies in customer relations, environmental hazards, seismology, and more Showcases successful applications in areas ranging from astronomy and the humanities to transport engineering Includes sample program snippets throughout the text as well as additional materials on a companion website The Data Bonanza is a must-have guide for information strategists, data analysts, and engineers in business, research, and government, and for anyone wishing to be on the cutting edge of data mining, machine learning, databases, distributed systems, or large-scale computing.

  3. Audio stream classification for multimedia database search

    NASA Astrophysics Data System (ADS)

    Artese, M.; Bianco, S.; Gagliardi, I.; Gasparini, F.

    2013-03-01

    Search and retrieval of huge archives of Multimedia data is a challenging task. A classification step is often used to reduce the number of entries on which to perform the subsequent search. In particular, when new entries of the database are continuously added, a fast classification based on simple threshold evaluation is desirable. In this work we present a CART-based (Classification And Regression Tree [1]) classification framework for audio streams belonging to multimedia databases. The database considered is the Archive of Ethnography and Social History (AESS) [2], which is mainly composed of popular songs and other audio records describing the popular traditions handed down generation by generation, such as traditional fairs, and customs. The peculiarities of this database are that it is continuously updated; the audio recordings are acquired in unconstrained environment; and for the non-expert human user is difficult to create the ground truth labels. In our experiments, half of all the available audio files have been randomly extracted and used as training set. The remaining ones have been used as test set. The classifier has been trained to distinguish among three different classes: speech, music, and song. All the audio files in the dataset have been previously manually labeled into the three classes above defined by domain experts.

  4. An Audio-Visual Tutorial Laboratory Program for Introductory Geology. Final Report.

    ERIC Educational Resources Information Center

    Sweet, Walter C.; Bates, Robert L.

    Described is an audio-visual tutorial laboratory program designed to provide a uniform, regularly reinforced, programmatic introduction to a limited set of geologic concepts and features; to provide a sequence of problem-solving exercises on which the student can work as an individual and in which he is required repeatedly to use all elements of…

  5. Transitioning from Analog to Digital Audio Recording in Childhood Speech Sound Disorders

    ERIC Educational Resources Information Center

    Shriberg, Lawrence D.; Mcsweeny, Jane L.; Anderson, Bruce E.; Campbell, Thomas F.; Chial, Michael R.; Green, Jordan R.; Hauner, Katherina K.; Moore, Christopher A.; Rusiewicz, Heather L.; Wilson, David L.

    2005-01-01

    Few empirical findings or technical guidelines are available on the current transition from analog to digital audio recording in childhood speech sound disorders. Of particular concern in the present context was whether a transition from analog- to digital-based transcription and coding of prosody and voice features might require re-standardizing…

  6. High performance MPEG-audio decoder IC

    NASA Technical Reports Server (NTRS)

    Thorn, M.; Benbassat, G.; Cyr, K.; Li, S.; Gill, M.; Kam, D.; Walker, K.; Look, P.; Eldridge, C.; Ng, P.

    1993-01-01

    The emerging digital audio and video compression technology brings both an opportunity and a new challenge to IC design. The pervasive application of compression technology to consumer electronics will require high volume, low cost IC's and fast time to market of the prototypes and production units. At the same time, the algorithms used in the compression technology result in complex VLSI IC's. The conflicting challenges of algorithm complexity, low cost, and fast time to market have an impact on device architecture and design methodology. The work presented in this paper is about the design of a dedicated, high precision, Motion Picture Expert Group (MPEG) audio decoder.

  7. Free audio archive brings legends to life

    NASA Astrophysics Data System (ADS)

    Banks, Michael

    2009-09-01

    A free online archive that contains thousands of hours of interviews with physicists and astronomers has been launched by the American Institute of Physics (AIP). The archive currently contains full transcripts of interviews with over 400 physicists that were recorded by historians and journalists from 1960 onwards as well as selected parts of the audio files. By the end of the year, the archive, belonging to the AIP's Niels Bohr Library and Archives in Washington, DC, will contain 500 online transcripts and over a dozen audio interviews.

  8. Nonlinear dynamic macromodeling techniques for audio systems

    NASA Astrophysics Data System (ADS)

    Ogrodzki, Jan; Bieńkowski, Piotr

    2015-09-01

    This paper develops a modelling method and a models identification technique for the nonlinear dynamic audio systems. Identification is performed by means of a behavioral approach based on a polynomial approximation. This approach makes use of Discrete Fourier Transform and Harmonic Balance Method. A model of an audio system is first created and identified and then it is simulated in real time using an algorithm of low computational complexity. The algorithm consists in real time emulation of the system response rather than in simulation of the system itself. The proposed software is written in Python language using object oriented programming techniques. The code is optimized for a multithreads environment.

  9. Hierarchical structure for audio-video based semantic classification of sports video sequences

    NASA Astrophysics Data System (ADS)

    Kolekar, M. H.; Sengupta, S.

    2005-07-01

    A hierarchical structure for sports event classification based on audio and video content analysis is proposed in this paper. Compared to the event classifications in other games, those of cricket are very challenging and yet unexplored. We have successfully solved cricket video classification problem using a six level hierarchical structure. The first level performs event detection based on audio energy and Zero Crossing Rate (ZCR) of short-time audio signal. In the subsequent levels, we classify the events based on video features using a Hidden Markov Model implemented through Dynamic Programming (HMM-DP) using color or motion as a likelihood function. For some of the game-specific decisions, a rule-based classification is also performed. Our proposed hierarchical structure can easily be applied to any other sports. Our results are very promising and we have moved a step forward towards addressing semantic classification problems in general.

  10. Three-dimensional audio versus head-down traffic alert and collision avoidance system displays.

    PubMed

    Begault, D R; Pittman, M T

    1996-01-01

    The advantage of a head-up auditory display for situational awareness was evaluated in an experiment designed to measure and compare the acquisition time for capturing visual targets under two conditions: standard head-down Traffic Alert and Collision Avoidance System display and three-dimensional (3-D) audio Traffic Alert and Collision Avoidance System presentation. (The technology used for 3-D audio presentation allows a stereo headphone user to potentially localize a sound at any externalized position in 3-D auditory space). Ten commercial airline crews were tested under full-mission simulation conditions at the NASA-Ames Crew-Vehicle Systems Research Facility Advanced Concepts Flight Simulator. Scenario software generated targets corresponding to aircraft that activated a 3-D aural advisory (the head-up auditory condition) or a standard, visual-audio TCAS advisory (map display with monaural audio alert). Results showed a significant difference in target acquisition time between the two conditions, favoring the 3-D audio Traffic Alert and Collision Avoidance System condition by 500 ms. PMID:11539173

  11. A Comprehensive Insight into the Chemical Space and ADME Features of Small Molecule NS5A Inhibitors.

    PubMed

    Ivanenkov, Yan A; Veselov, Mark S; Shakhbazyan, Artem G; Aladinskiy, Vladimir A; Aladinskaya, Anastasia V; Yartseva, Sofya M; Majouga, Alexander G; Vantskul, Anton S; Leonov, Sergey V; Ivachtchenko, Alexandre V; Koteliansky, Victor E

    2016-01-01

    Non-structural 5A (NS5A) protein plays a crucial role in the replication of hepatitis C virus (HCV) and during the past decade has attracted increasing attention as a promising biological target for the treatment of viral infections and related disorders. Small-molecule NS5A inhibitors have shown significant antiviral activity in vitro and in vivo. Several lead molecules are reasonably regarded as novel highly potent drug candidates with favorable ADME features and tolerable side effects. The first-in-class daclatasvir has recently been launched into the market and 14 novel molecules are currently under evaluation in clinical trials. From this perspective, we provide an overview of the available chemical space of small-molecule NS5A inhibitors and their PK properties, mainly focusing on the diversity in structure and scaffold representation. PMID:26585933

  12. Key Features of Human Episodic Recollection in the Cross-Episode Retrieval of Rat Hippocampus Representations of Space

    PubMed Central

    Kelemen, Eduard; Fenton, André A.

    2013-01-01

    Neurophysiological studies focus on memory retrieval as a reproduction of what was experienced and have established that neural discharge is replayed to express memory. However, cognitive psychology has established that recollection is not a verbatim replay of stored information. Recollection is constructive, the product of memory retrieval cues, the information stored in memory, and the subject's state of mind. We discovered key features of constructive recollection embedded in the rat CA1 ensemble discharge during an active avoidance task. Rats learned two task variants, one with the arena stable, the other with it rotating; each variant defined a distinct behavioral episode. During the rotating episode, the ensemble discharge of CA1 principal neurons was dynamically organized to concurrently represent space in two distinct codes. The code for spatial reference frame switched rapidly between representing the rat's current location in either the stationary spatial frame of the room or the rotating frame of the arena. The code for task variant switched less frequently between a representation of the current rotating episode and the stable episode from the rat's past. The characteristics and interplay of these two hippocampal codes revealed three key properties of constructive recollection. (1) Although the ensemble representations of the stable and rotating episodes were distinct, ensemble discharge during rotation occasionally resembled the stable condition, demonstrating cross-episode retrieval of the representation of the remote, stable episode. (2) This cross-episode retrieval at the level of the code for task variant was more likely when the rotating arena was about to match its orientation in the stable episode. (3) The likelihood of cross-episode retrieval was influenced by preretrieval information that was signaled at the level of the code for spatial reference frame. Thus key features of episodic recollection manifest in rat hippocampal representations of

  13. Music information retrieval in compressed audio files: a survey

    NASA Astrophysics Data System (ADS)

    Zampoglou, Markos; Malamos, Athanasios G.

    2014-07-01

    In this paper, we present an organized survey of the existing literature on music information retrieval systems in which descriptor features are extracted directly from the compressed audio files, without prior decompression to pulse-code modulation format. Avoiding the decompression step and utilizing the readily available compressed-domain information can significantly lighten the computational cost of a music information retrieval system, allowing application to large-scale music databases. We identify a number of systems relying on compressed-domain information and form a systematic classification of the features they extract, the retrieval tasks they tackle and the degree in which they achieve an actual increase in the overall speed-as well as any resulting loss in accuracy. Finally, we discuss recent developments in the field, and the potential research directions they open toward ultra-fast, scalable systems.

  14. Cross-modal retrieval of scripted speech audio

    NASA Astrophysics Data System (ADS)

    Owen, Charles B.; Makedon, Fillia

    1997-12-01

    This paper describes an approach to the problem of searching speech-based digital audio using cross-modal information retrieval. Audio containing speech (speech-based audio) is difficult to search. Open vocabulary speech recognition is advancing rapidly, but cannot yield high accuracy in either search or transcription modalities. However, text can be searched quickly and efficiently with high accuracy. Script- light digital audio is audio that has an available transcription. This is a surprisingly large class of content including legal testimony, broadcasting, dramatic productions and political meetings and speeches. An automatic mechanism for deriving the synchronization between the transcription and the audio allows for very accurate retrieval of segments of that audio. The mechanism described in this paper is based on building a transcription graph from the text and computing biphone probabilities for the audio. A modified beam search algorithm is presented to compute the alignment.

  15. Safe, Affordable, Convenient: Environmental Features of Malls and Other Public Spaces Used by Older Adults for Walking

    PubMed Central

    King, Diane K.; Allen, Peg; Jones, Dina L.; Marquez, David X.; Brown, David R.; Rosenberg, Dori; Janicek, Sarah; Allen, Laila; Belza, Basia

    2016-01-01

    Background Midlife and older adults use shopping malls for walking, but little research has examined mall characteristics that contribute to their walkability. Methods We used modified versions of the Centers for Disease Control and Prevention (CDC)-Healthy Aging Research Network (HAN) Environmental Audit and the System for Observing Play and Recreation in Communities (SOPARC) tool to systematically observe 443 walkers in 10 shopping malls. We also observed 87 walkers in 6 community-based nonmall/nongym venues where older adults routinely walked for physical activity. Results All venues had public transit stops and accessible parking. All malls and 67% of nonmalls had wayfinding aids, and most venues (81%) had an established circuitous walking route and clean, well-maintained public restrooms (94%). All venues had level floor surfaces, and one-half had benches along the walking route. Venues varied in hours of access, programming, tripping hazards, traffic control near entrances, and lighting. Conclusions Despite diversity in location, size, and purpose, the mall and nonmall venues audited shared numerous environmental features known to promote walking in older adults and few barriers to walking. Future research should consider programmatic features and outreach strategies to expand the use of malls and other suitable public spaces for walking. PMID:26181907

  16. Lip-Reading Aids Word Recognition Most in Moderate Noise: A Bayesian Explanation Using High-Dimensional Feature Space

    PubMed Central

    Ross, Lars A.; Foxe, John J.; Parra, Lucas C.

    2009-01-01

    Watching a speaker's facial movements can dramatically enhance our ability to comprehend words, especially in noisy environments. From a general doctrine of combining information from different sensory modalities (the principle of inverse effectiveness), one would expect that the visual signals would be most effective at the highest levels of auditory noise. In contrast, we find, in accord with a recent paper, that visual information improves performance more at intermediate levels of auditory noise than at the highest levels, and we show that a novel visual stimulus containing only temporal information does the same. We present a Bayesian model of optimal cue integration that can explain these conflicts. In this model, words are regarded as points in a multidimensional space and word recognition is a probabilistic inference process. When the dimensionality of the feature space is low, the Bayesian model predicts inverse effectiveness; when the dimensionality is high, the enhancement is maximal at intermediate auditory noise levels. When the auditory and visual stimuli differ slightly in high noise, the model makes a counterintuitive prediction: as sound quality increases, the proportion of reported words corresponding to the visual stimulus should first increase and then decrease. We confirm this prediction in a behavioral experiment. We conclude that auditory-visual speech perception obeys the same notion of optimality previously observed only for simple multisensory stimuli. PMID:19259259

  17. Audio/Visual Ratios in Commercial Filmstrips.

    ERIC Educational Resources Information Center

    Gulliford, Nancy L.

    Developed by the Westinghouse Electric Corporation, Video Audio Compressed (VIDAC) is a compressed time, variable rate, still picture television system. This technology made it possible for a centralized library of audiovisual materials to be transmitted over a television channel in very short periods of time. In order to establish specifications…

  18. Building Digital Audio Preservation Infrastructure and Workflows

    ERIC Educational Resources Information Center

    Young, Anjanette; Olivieri, Blynne; Eckler, Karl; Gerontakos, Theodore

    2010-01-01

    In 2009 the University of Washington (UW) Libraries special collections received funding for the digital preservation of its audio indigenous language holdings. The university libraries, where the authors work in various capacities, had begun digitizing image and text collections in 1997. Because of this, at the onset of the project, workflows (a…

  19. An ESL Audio-Script Writing Workshop

    ERIC Educational Resources Information Center

    Miller, Carla

    2012-01-01

    The roles of dialogue, collaborative writing, and authentic communication have been explored as effective strategies in second language writing classrooms. In this article, the stages of an innovative, multi-skill writing method, which embeds students' personal voices into the writing process, are explored. A 10-step ESL Audio Script Writing Model…

  20. Providing Students with Formative Audio Feedback

    ERIC Educational Resources Information Center

    Brearley, Francis Q.; Cullen, W. Rod

    2012-01-01

    The provision of timely and constructive feedback is increasingly challenging for busy academics. Ensuring effective student engagement with feedback is equally difficult. Increasingly, studies have explored provision of audio recorded feedback to enhance effectiveness and engagement with feedback. Few, if any, of these focus on purely formative…

  1. Spanish for Agricultural Purposes: The Audio Program.

    ERIC Educational Resources Information Center

    Mainous, Bruce H.; And Others

    The manual is meant to accompany and supplement the basic manual and to serve as support to the audio component of "Spanish for Agricultural Purposes," a one-semester course for North American agriculture specialists preparing to work in Latin America, consists of exercises to supplement readings presented in the course's basic manual and to…

  2. Structuring Broadcast Audio for Information Access

    NASA Astrophysics Data System (ADS)

    Gauvain, Jean-Luc; Lamel, Lori

    2003-12-01

    One rapidly expanding application area for state-of-the-art speech recognition technology is the automatic processing of broadcast audiovisual data for information access. Since much of the linguistic information is found in the audio channel, speech recognition is a key enabling technology which, when combined with information retrieval techniques, can be used for searching large audiovisual document collections. Audio indexing must take into account the specificities of audio data such as needing to deal with the continuous data stream and an imperfect word transcription. Other important considerations are dealing with language specificities and facilitating language portability. At Laboratoire d'Informatique pour la Mécanique et les Sciences de l'Ingénieur (LIMSI), broadcast news transcription systems have been developed for seven languages: English, French, German, Mandarin, Portuguese, Spanish, and Arabic. The transcription systems have been integrated into prototype demonstrators for several application areas such as audio data mining, structuring audiovisual archives, selective dissemination of information, and topic tracking for media monitoring. As examples, this paper addresses the spoken document retrieval and topic tracking tasks.

  3. Solar Energy Audio-Visual Materials.

    ERIC Educational Resources Information Center

    Department of Housing and Urban Development, Washington, DC. Office of Policy Development and Research.

    This directory presents an annotated bibliography of non-print information resources dealing with solar energy. The document is divided by type of audio-visual medium, including: (1) Films, (2) Slides and Filmstrips, and (3) Videotapes. A fourth section provides addresses and telephone numbers of audiovisual aids sources, and lists the page…

  4. 50 CFR 27.72 - Audio equipment.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 50 Wildlife and Fisheries 8 2011-10-01 2011-10-01 false Audio equipment. 27.72 Section 27.72 Wildlife and Fisheries UNITED STATES FISH AND WILDLIFE SERVICE, DEPARTMENT OF THE INTERIOR (CONTINUED) THE NATIONAL WILDLIFE REFUGE SYSTEM PROHIBITED ACTS Disturbing Violations: Light and Sound Equipment §...

  5. 50 CFR 27.72 - Audio equipment.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 50 Wildlife and Fisheries 9 2012-10-01 2012-10-01 false Audio equipment. 27.72 Section 27.72 Wildlife and Fisheries UNITED STATES FISH AND WILDLIFE SERVICE, DEPARTMENT OF THE INTERIOR (CONTINUED) THE NATIONAL WILDLIFE REFUGE SYSTEM PROHIBITED ACTS Disturbing Violations: Light and Sound Equipment §...

  6. 50 CFR 27.72 - Audio equipment.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 50 Wildlife and Fisheries 9 2013-10-01 2013-10-01 false Audio equipment. 27.72 Section 27.72 Wildlife and Fisheries UNITED STATES FISH AND WILDLIFE SERVICE, DEPARTMENT OF THE INTERIOR (CONTINUED) THE NATIONAL WILDLIFE REFUGE SYSTEM PROHIBITED ACTS Disturbing Violations: Filming, Photography, and Light...

  7. 50 CFR 27.72 - Audio equipment.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 50 Wildlife and Fisheries 9 2014-10-01 2014-10-01 false Audio equipment. 27.72 Section 27.72 Wildlife and Fisheries UNITED STATES FISH AND WILDLIFE SERVICE, DEPARTMENT OF THE INTERIOR (CONTINUED) THE NATIONAL WILDLIFE REFUGE SYSTEM PROHIBITED ACTS Disturbing Violations: Filming, Photography, and Light...

  8. Improving Audio Quality in Distance Learning Applications.

    ERIC Educational Resources Information Center

    Richardson, Craig H.

    This paper discusses common causes of problems encountered with audio systems in distance learning networks and offers practical suggestions for correcting the problems. Problems and discussions are divided into nine categories: (1) acoustics, including reverberant classrooms leading to distorted or garbled voices, as well as one-dimensional audio…

  9. 47 CFR 10.520 - Common audio attention signal.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 47 Telecommunication 1 2013-10-01 2013-10-01 false Common audio attention signal. 10.520 Section... Equipment Requirements § 10.520 Common audio attention signal. A Participating CMS Provider and equipment manufacturers may only market devices for public use under part 10 that include an audio attention signal...

  10. 47 CFR 10.520 - Common audio attention signal.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 47 Telecommunication 1 2012-10-01 2012-10-01 false Common audio attention signal. 10.520 Section... Equipment Requirements § 10.520 Common audio attention signal. A Participating CMS Provider and equipment manufacturers may only market devices for public use under part 10 that include an audio attention signal...

  11. 47 CFR 87.483 - Audio visual warning systems.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 47 Telecommunication 5 2014-10-01 2014-10-01 false Audio visual warning systems. 87.483 Section 87... AVIATION SERVICES Stations in the Radiodetermination Service § 87.483 Audio visual warning systems. An audio visual warning system (AVWS) is a radar-based obstacle avoidance system. AVWS...

  12. 47 CFR 10.520 - Common audio attention signal.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 47 Telecommunication 1 2014-10-01 2014-10-01 false Common audio attention signal. 10.520 Section... Equipment Requirements § 10.520 Common audio attention signal. A Participating CMS Provider and equipment manufacturers may only market devices for public use under part 10 that include an audio attention signal...

  13. 47 CFR 10.520 - Common audio attention signal.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 47 Telecommunication 1 2010-10-01 2010-10-01 false Common audio attention signal. 10.520 Section... Equipment Requirements § 10.520 Common audio attention signal. A Participating CMS Provider and equipment manufacturers may only market devices for public use under part 10 that include an audio attention signal...

  14. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 47 Telecommunication 4 2011-10-01 2011-10-01 false Digital audio broadcasting service requirements... SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast radio stations using IBOC must transmit at least one over-the-air digital...

  15. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 47 Telecommunication 4 2014-10-01 2014-10-01 false Digital audio broadcasting service requirements... SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast radio stations using IBOC must transmit at least one over-the-air digital...

  16. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 47 Telecommunication 4 2013-10-01 2013-10-01 false Digital audio broadcasting service requirements. 73.403 Section 73.403 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST RADIO SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast...

  17. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 47 Telecommunication 4 2012-10-01 2012-10-01 false Digital audio broadcasting service requirements. 73.403 Section 73.403 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST RADIO SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast...

  18. To Make a Long Story Short: Abridged Audio at 10.

    ERIC Educational Resources Information Center

    Annichiarico, Mark

    1996-01-01

    Examines the history of abridged audio publishing 10 years after the formation of the Audio Publishers Association. Topics include abridged versus unabridged versions for bookstores and libraries; vendors and publishers; future possibilities for CDs and DVD (Digital Versatile Disc); and audio leasing for libraries. (LRW)

  19. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... programming stream at no direct charge to listeners. In addition, a broadcast radio station must simulcast its analog audio programming on one of its digital audio programming streams. The DAB audio programming... analog programming service currently provided to listeners. (b) Emergency information. The...

  20. 43 CFR 8365.2-2 - Audio devices.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 43 Public Lands: Interior 2 2011-10-01 2011-10-01 false Audio devices. 8365.2-2 Section 8365.2-2..., DEPARTMENT OF THE INTERIOR RECREATION PROGRAMS VISITOR SERVICES Rules of Conduct § 8365.2-2 Audio devices. On... audio device such as a radio, television, musical instrument, or other noise producing device...

  1. Internet-oriented visualization with audio presentation of speech signals

    NASA Astrophysics Data System (ADS)

    Braun, Jerome J.; Levkowitz, Haim

    1998-05-01

    Visualization of speech signals, including the capability to visualize the waveforms while simultaneously hearing the speech, is among the essential requirements in speech processing research. In tasks related to labeling of speech signals, visualization activities may have to be performed by multiple users upon a centralized collection of speech data. When speech labeling activities involve perceptual issues, the human factors issues including functionality tradeoffs are particularly important, since the user's burden (tiredness, annoyance) can affect the perceptual responses. We developed VideVox (pronounced 'Veedeh-Vox'), a speech visualization facility, in which the visualization activities may be performed by a large number of users in geographically, dialectically and linguistically diverse locations. Developed in Java, and capable of operating both as an Internet Java applet and a Java application, VideVox is platform independent. Using the client-server architecture paradigm, it allows distributed visualization work. The Internet orientation makes VideVox a promising direction for speech signal visualization in speech labeling activities that require a large number of users in multiple locations. In the paper, we describe our approach, VideVox features, modes of audio data exploration and audio-synchronous animation for speech visualization, operations related to identification of perceptual events, and the human factors issues related to perception-oriented visualizations of speech.

  2. Head Tracking of Auditory, Visual, and Audio-Visual Targets.

    PubMed

    Leung, Johahn; Wei, Vincent; Burgess, Martin; Carlile, Simon

    2015-01-01

    The ability to actively follow a moving auditory target with our heads remains unexplored even though it is a common behavioral response. Previous studies of auditory motion perception have focused on the condition where the subjects are passive. The current study examined head tracking behavior to a moving auditory target along a horizontal 100° arc in the frontal hemisphere, with velocities ranging from 20 to 110°/s. By integrating high fidelity virtual auditory space with a high-speed visual presentation we compared tracking responses of auditory targets against visual-only and audio-visual "bisensory" stimuli. Three metrics were measured-onset, RMS, and gain error. The results showed that tracking accuracy (RMS error) varied linearly with target velocity, with a significantly higher rate in audition. Also, when the target moved faster than 80°/s, onset and RMS error were significantly worst in audition the other modalities while responses in the visual and bisensory conditions were statistically identical for all metrics measured. Lastly, audio-visual facilitation was not observed when tracking bisensory targets. PMID:26778952

  3. Head Tracking of Auditory, Visual, and Audio-Visual Targets

    PubMed Central

    Leung, Johahn; Wei, Vincent; Burgess, Martin; Carlile, Simon

    2016-01-01

    The ability to actively follow a moving auditory target with our heads remains unexplored even though it is a common behavioral response. Previous studies of auditory motion perception have focused on the condition where the subjects are passive. The current study examined head tracking behavior to a moving auditory target along a horizontal 100° arc in the frontal hemisphere, with velocities ranging from 20 to 110°/s. By integrating high fidelity virtual auditory space with a high-speed visual presentation we compared tracking responses of auditory targets against visual-only and audio-visual “bisensory” stimuli. Three metrics were measured—onset, RMS, and gain error. The results showed that tracking accuracy (RMS error) varied linearly with target velocity, with a significantly higher rate in audition. Also, when the target moved faster than 80°/s, onset and RMS error were significantly worst in audition the other modalities while responses in the visual and bisensory conditions were statistically identical for all metrics measured. Lastly, audio-visual facilitation was not observed when tracking bisensory targets. PMID:26778952

  4. Segmentation of perivascular spaces in 7T MR image using auto-context model with orientation-normalized features.

    PubMed

    Park, Sang Hyun; Zong, Xiaopeng; Gao, Yaozong; Lin, Weili; Shen, Dinggang

    2016-07-01

    Quantitative study of perivascular spaces (PVSs) in brain magnetic resonance (MR) images is important for understanding the brain lymphatic system and its relationship with neurological diseases. One of the major challenges is the accurate extraction of PVSs that have very thin tubular structures with various directions in three-dimensional (3D) MR images. In this paper, we propose a learning-based PVS segmentation method to address this challenge. Specifically, we first determine a region of interest (ROI) by using the anatomical brain structure and the vesselness information derived from eigenvalues of image derivatives. Then, in the ROI, we extract a number of randomized Haar features which are normalized with respect to the principal directions of the underlying image derivatives. The classifier is trained by the random forest model that can effectively learn both discriminative features and classifier parameters to maximize the information gain. Finally, a sequential learning strategy is used to further enforce various contextual patterns around the thin tubular structures into the classifier. For evaluation, we apply our proposed method to the 7T brain MR images scanned from 17 healthy subjects aged from 25 to 37. The performance is measured by voxel-wise segmentation accuracy, cluster-wise classification accuracy, and similarity of geometric properties, such as volume, length, and diameter distributions between the predicted and the true PVSs. Moreover, the accuracies are also evaluated on the simulation images with motion artifacts and lacunes to demonstrate the potential of our method in segmenting PVSs from elderly and patient populations. The experimental results show that our proposed method outperforms all existing PVS segmentation methods. PMID:27046107

  5. Three-dimensional audio using loudspeakers

    NASA Astrophysics Data System (ADS)

    Gardner, William G.

    1997-12-01

    3-D audio systems, which can surround a listener with sounds at arbitrary locations, are an important part of immersive interfaces. A new approach is presented for implementing 3-D audio using a pair of conventional loudspeakers. The new idea is to use the tracked position of the listener's head to optimize the acoustical presentation, and thus produce a much more realistic illusion over a larger listening area than existing loudspeaker 3-D audio systems. By using a remote head tracker, for instance based on computer vision, an immersive audio environment can be created without donning headphones or other equipment. The general approach to a 3-D audio system is to reconstruct the acoustic pressures at the listener's ears that would result from the natural listening situation to be simulated. To accomplish this using loudspeakers requires that first, the ear signals corresponding to the target scene are synthesized by appropriately encoding directional cues, a process known as 'binaural synthesis,' and second, these signals are delivered to the listener by inverting the transmission paths that exist from the speakers to the listener, a process known as 'crosstalk cancellation.' Existing crosstalk cancellation systems only function at a fixed listening location; when the listener moves away from the equalization zone, the 3-D illusion is lost. Steering the equalization zone to the tracked listener preserves the 3-D illusion over a large listening volume, thus simulating a reconstructed soundfield, and also provides dynamic localization cues by maintaining stationary external sound sources during head motion. This dissertation will discuss the theory, implementation, and testing of a head-tracked loudspeaker 3-D audio system. Crosstalk cancellers that can be steered to the location of a tracked listener will be described. The objective performance of these systems has been evaluated using simulations and acoustical measurements made at the ears of human subjects. Many

  6. Comparing Audio and Video Data for Rating Communication

    PubMed Central

    Williams, Kristine; Herman, Ruth; Bontempo, Daniel

    2013-01-01

    Video recording has become increasingly popular in nursing research, adding rich nonverbal, contextual, and behavioral information. However, benefits of video over audio data have not been well established. We compared communication ratings of audio versus video data using the Emotional Tone Rating Scale. Twenty raters watched video clips of nursing care and rated staff communication on 12 descriptors that reflect dimensions of person-centered and controlling communication. Another group rated audio-only versions of the same clips. Interrater consistency was high within each group with ICC (2,1) for audio = .91, and video = .94. Interrater consistency for both groups combined was also high with ICC (2,1) for audio and video = .95. Communication ratings using audio and video data were highly correlated. The value of video being superior to audio recorded data should be evaluated in designing studies evaluating nursing care. PMID:23579475

  7. Demonstration of pattern transfer into sub-100 nm polysilicon line/space features patterned with extreme ultraviolet lithography

    SciTech Connect

    Cardinale, G. F.; Henderson, C. C.; Goldsmith, J. E. M.; Mangat, P. J. S.; Cobb, J.; Hector, S. D.

    1999-11-01

    In two separate experiments, we have successfully demonstrated the transfer of dense- and loose-pitch line/space (L/S) photoresist features, patterned with extreme ultraviolet (EUV) lithography, into an underlying hard mask material. In both experiments, a deep-UV photoresist ({approx}90 nm thick) was spin cast in bilayer format onto a hard mask (50-90 nm thick) and was subsequently exposed to EUV radiation using a 10x reduction EUV exposure system. The EUV reticle was fabricated at Motorola (Tempe, AZ) using a subtractive process with Ta-based absorbers on Mo/Si multilayer mask blanks. In the first set of experiments, following the EUV exposures, the L/S patterns were transferred first into a SiO{sub 2} hard mask (60 nm thick) using a reactive ion etch (RIE), and then into polysilicon (350 nm thick) using a triode-coupled plasma RIE etcher at the University of California, Berkeley, microfabrication facilities. The latter etch process, which produced steep (>85 degree sign ) sidewalls, employed a HBr/Cl chemistry with a large (>10:1) etch selectivity of polysilicon to silicon dioxide. In the second set of experiments, hard mask films of SiON (50 nm thick) and SiO{sub 2} (87 nm thick) were used. A RIE was performed at Motorola using a halogen gas chemistry that resulted in a hard mask-to-photoresist etch selectivity >3:1 and sidewall profile angles {>=}85 degree sign . Line edge roughness (LER) and linewidth critical dimension (CD) measurements were performed using Sandia's GORA(c) CD digital image analysis software. Low LER values (6-9 nm, 3{sigma}, one side) and good CD linearity (better than 10%) were demonstrated for the final pattern-transferred dense polysilicon L/S features from 80 to 175 nm. In addition, pattern transfer (into polysilicon) of loose-pitch (1:2) L/S features with CDs{>=}60 nm was demonstrated. (c) 1999 American Vacuum Society.

  8. Motif distributions in phase-space networks for characterizing experimental two-phase flow patterns with chaotic features.

    PubMed

    Gao, Zhong-Ke; Jin, Ning-De; Wang, Wen-Xu; Lai, Ying-Cheng

    2010-07-01

    The dynamics of two-phase flows have been a challenging problem in nonlinear dynamics and fluid mechanics. We propose a method to characterize and distinguish patterns from inclined water-oil flow experiments based on the concept of network motifs that have found great usage in network science and systems biology. In particular, we construct from measured time series phase-space complex networks and then calculate the distribution of a set of distinct network motifs. To gain insight, we first test the approach using time series from classical chaotic systems and find a universal feature: motif distributions from different chaotic systems are generally highly heterogeneous. Our main finding is that the distributions from experimental two-phase flows tend to be heterogeneous as well, suggesting the underlying chaotic nature of the flow patterns. Calculation of the maximal Lyapunov exponent provides further support for this. Motif distributions can thus be a feasible tool to understand the dynamics of realistic two-phase flow patterns. PMID:20866710

  9. Audio and video bit-rate reduction

    NASA Astrophysics Data System (ADS)

    Haskell, B. G.; Steele, R.

    1981-02-01

    The problem of estimating limits of bit-rate reduction for audio and video information is addressed. It is noted that binary encoding of such signals is performed by a variety of techniques, which are classified here into two categories (waveform and parameter). Encoders are compared on the basis of bits per Nyquist sample. Since minimizing the Nyquist sample requires a use of source signal statistics and human perception, audio and video sources and their statistics are discussed, as are the salient properties of hearing and vision and subjective assessments of quality. The present state of waveform and parameter encoding is assessed. It is predicted that waveform encoders will be refined to the point where bit rates will be nearly optimum but also highly variable. It is concluded that dramatic reductions will come about only through sophisticated parameter encoding.

  10. Capacity-optimized mp2 audio watermarking

    NASA Astrophysics Data System (ADS)

    Steinebach, Martin; Dittmann, Jana

    2003-06-01

    Today a number of audio watermarking algorithms have been proposed, some of them at a quality making them suitable for commercial applications. The focus of most of these algorithms is copyright protection. Therefore, transparency and robustness are the most discussed and optimised parameters. But other applications for audio watermarking can also be identified stressing other parameters like complexity or payload. In our paper, we introduce a new mp2 audio watermarking algorithm optimised for high payload. Our algorithm uses the scale factors of an mp2 file for watermark embedding. They are grouped and masked based on a pseudo-random pattern generated from a secret key. In each group, we embed one bit. Depending on the bit to embed, we change the scale factors by adding 1 where necessary until it includes either more even or uneven scale factors. An uneven group has a 1 embedded, an even group a 0. The same rule is later applied to detect the watermark. The group size can be increased or decreased for transparency/payload trade-off. We embed 160 bits or more in an mp2 file per second without reducing perceived quality. As an application example, we introduce a prototypic Karaoke system displaying song lyrics embedded as a watermark.

  11. Perceptually controlled doping for audio source separation

    NASA Astrophysics Data System (ADS)

    Mahé, Gaël; Nadalin, Everton Z.; Suyama, Ricardo; Romano, João MT

    2014-12-01

    The separation of an underdetermined audio mixture can be performed through sparse component analysis (SCA) that relies however on the strong hypothesis that source signals are sparse in some domain. To overcome this difficulty in the case where the original sources are available before the mixing process, the informed source separation (ISS) embeds in the mixture a watermark, which information can help a further separation. Though powerful, this technique is generally specific to a particular mixing setup and may be compromised by an additional bitrate compression stage. Thus, instead of watermarking, we propose a `doping' method that makes the time-frequency representation of each source more sparse, while preserving its audio quality. This method is based on an iterative decrease of the distance between the distribution of the signal and a target sparse distribution, under a perceptual constraint. We aim to show that the proposed approach is robust to audio coding and that the use of the sparsified signals improves the source separation, in comparison with the original sources. In this work, the analysis is made only in instantaneous mixtures and focused on voice sources.

  12. An active headrest for personal audio.

    PubMed

    Elliott, Stephen J; Jones, Matthew

    2006-05-01

    There is an increasing need for personal audio systems, which generate sounds that are clearly audible to one listener but are not audible to other listeners nearby. Of particular interest in this paper are listeners sitting in adjacent seats in aircraft or land vehicles. Although personal audio could then be achieved with headsets, it would be safer and more comfortable if loudspeakers in the seat headrests could be actively controlled to generate an acceptable level of acoustic isolation. In this paper a number of approaches to this problem are investigated, but the most successful involves a pair of loudspeakers on one side of the headrest, driven together to reproduce an audio signal for a listener in that seat and also to attenuate the pressures in the adjacent seat. The performance of this technique is investigated using simple analytic models and also with a practical implementation, tested in an anechoic chamber and a small room. It is found that significant attenuations, of between 5 and 25 dB, can be achieved in the crosstalk between the seats for frequencies up to about 2 kHz. PMID:16708929

  13. Feature-space assessment of electrical impedance tomography coregistered with computed tomography in detecting multiple contrast targets

    SciTech Connect

    Krishnan, Kalpagam; Liu, Jeff; Kohli, Kirpal

    2014-06-15

    Purpose: Fusion of electrical impedance tomography (EIT) with computed tomography (CT) can be useful as a clinical tool for providing additional physiological information about tissues, but requires suitable fusion algorithms and validation procedures. This work explores the feasibility of fusing EIT and CT images using an algorithm for coregistration. The imaging performance is validated through feature space assessment on phantom contrast targets. Methods: EIT data were acquired by scanning a phantom using a circuit, configured for injecting current through 16 electrodes, placed around the phantom. A conductivity image of the phantom was obtained from the data using electrical impedance and diffuse optical tomography reconstruction software (EIDORS). A CT image of the phantom was also acquired. The EIT and CT images were fused using a region of interest (ROI) coregistration fusion algorithm. Phantom imaging experiments were carried out on objects of different contrasts, sizes, and positions. The conductive medium of the phantoms was made of a tissue-mimicking bolus material that is routinely used in clinical radiation therapy settings. To validate the imaging performance in detecting different contrasts, the ROI of the phantom was filled with distilled water and normal saline. Spatially separated cylindrical objects of different sizes were used for validating the imaging performance in multiple target detection. Analyses of the CT, EIT and the EIT/CT phantom images were carried out based on the variations of contrast, correlation, energy, and homogeneity, using a gray level co-occurrence matrix (GLCM). A reference image of the phantom was simulated using EIDORS, and the performances of the CT and EIT imaging systems were evaluated and compared against the performance of the EIT/CT system using various feature metrics, detectability, and structural similarity index measures. Results: In detecting distilled and normal saline water in bolus medium, EIT as a stand

  14. High-performance combination method of electric network frequency and phase for audio forgery detection in battery-powered devices.

    PubMed

    Savari, Maryam; Abdul Wahab, Ainuddin Wahid; Anuar, Nor Badrul

    2016-09-01

    Audio forgery is any act of tampering, illegal copy and fake quality in the audio in a criminal way. In the last decade, there has been increasing attention to the audio forgery detection due to a significant increase in the number of forge in different type of audio. There are a number of methods for forgery detection, which electric network frequency (ENF) is one of the powerful methods in this area for forgery detection in terms of accuracy. In spite of suitable accuracy of ENF in a majority of plug-in powered devices, the weak accuracy of ENF in audio forgery detection for battery-powered devices, especially in laptop and mobile phone, can be consider as one of the main obstacles of the ENF. To solve the ENF problem in terms of accuracy in battery-powered devices, a combination method of ENF and phase feature is proposed. From experiment conducted, ENF alone give 50% and 60% accuracy for forgery detection in mobile phone and laptop respectively, while the proposed method shows 88% and 92% accuracy respectively, for forgery detection in battery-powered devices. The results lead to higher accuracy for forgery detection with the combination of ENF and phase feature. PMID:27442454

  15. A digital audio/video interleaving system. [for Shuttle Orbiter

    NASA Technical Reports Server (NTRS)

    Richards, R. W.

    1978-01-01

    A method of interleaving an audio signal with its associated video signal for simultaneous transmission or recording, and the subsequent separation of the two signals, is described. Comparisons are made between the new audio signal interleaving system and the Skylab Pam audio/video interleaving system, pointing out improvements gained by using the digital audio/video interleaving system. It was found that the digital technique is the simplest, most effective and most reliable method for interleaving audio and/or other types of data into the video signal for the Shuttle Orbiter application. Details of the design of a multiplexer capable of accommodating two basic data channels, each consisting of a single 31.5-kb/s digital bit stream are given. An adaptive slope delta modulation system is introduced to digitize audio signals, producing a high immunity of work intelligibility to channel errors, primarily due to the robust nature of the delta-modulation algorithm.

  16. Objective child behavior measurement with naturalistic daylong audio recording and its application to autism identification.

    PubMed

    Xu, Dongxin; Gilkerson, Jill; Richards, Jeffrey A

    2012-01-01

    Child behavior in the natural environment is a subject that is relevant for many areas of social science and bio-behavioral research. However, its measurement is currently based mainly on subjective approaches such as parent questionnaires or clinical observation. This study demonstrates an objective and unobtrusive child vocal behavior measurement and monitoring approach using daylong audio recordings of children in the natural home environment. Our previous research has shown significant performance in childhood autism identification. However, there remains the question of why it works. In the previous study, the focus was more on the overall performance and data-driven modeling without regard to the meaning of underlying features. Even if a high risk of autism is predicted, specific information about child behavior that could contribute to the automated categorization was not further explored. This study attempts to clarify this issue by exploring the details of underlying features and uncovering additional behavioral information buried within the audio streams. It was found that much child vocal behavior can be measured automatically by applying signal processing and pattern recognition technologies to daylong audio recordings. By combining many such features, the model achieves an overall autism identification accuracy of 94% (N=226). Similar to many emerging non-invasive and telemonitoring technologies in health care, this approach is believed to have great potential in child development research, clinical practice and parenting. PMID:23366733

  17. A content-based digital audio watermarking algorithm

    NASA Astrophysics Data System (ADS)

    Zhang, Liping; Zhao, Yi; Xu, Wen Li

    2015-12-01

    Digital audio watermarking embeds inaudible information into digital audio data for the purposes of copyright protection, ownership verification, covert communication, and/or auxiliary data carrying. In this paper, we present a novel watermarking scheme to embed a meaningful gray image into digital audio by quantizing the wavelet coefficients (using integer lifting wavelet transform) of audio samples. Our audio-dependent watermarking procedure directly exploits temporal and frequency perceptual masking of the human auditory system (HAS) to guarantee that the embedded watermark image is inaudible and robust. The watermark is constructed by utilizing still image compression technique, breaking each audio clip into smaller segments, selecting the perceptually significant audio segments to wavelet transform, and quantizing the perceptually significant wavelet coefficients. The proposed watermarking algorithm can extract the watermark image without the help from the original digital audio signals. We also demonstrate the robustness of that watermarking procedure to audio degradations and distortions, e.g., those that result from noise adding, MPEG compression, low pass filtering, resampling, and requantization.

  18. Features of motivation of the crewmembers in an enclosed space at atmospheric pressure changes during breathing inert gases.

    NASA Astrophysics Data System (ADS)

    Komarevcev, Sergey

    Since the 1960s, our psychologists are working on experimenting with small groups in isolation .It was associated with the beginning of spaceflight and necessity to study of human behaviors in ways different from the natural habitat of man .Those, who study human behavior especially in isolation, know- that the behavior in isolation markedly different from that in the natural situаtions. It associated with the development of new, more adaptive behaviors (1) What are the differences ? First of all , isolation is achieved by the fact ,that the group is in a closed space. How experiments show - the crew members have changed the basic personality traits, such as motivation Statement of the problem and methods. In our experimentation we were interested in changing the features of human motivation (strength, stability and direction of motivation) in terms of a closed group in the modified atmosphere pressure and breathing inert gases. Also, we were interested in particular external and internal motivation of the individual in the circumstances. To conduct experimentation , we used an experimental barocomplex GVK -250 , which placed a group of six mаns. A task was to spend fifteen days in isolation on barokomplex when breathing oxigen - xenon mixture of fifteen days in isolation on the same complex when breathing oxygen- helium mixture and fifteen days of isolation on the same complex when breathing normal air All this time, the subjects were isolated under conditions of atmospheric pressure changes , closer to what you normally deal divers. We assumed that breathing inert mixtures can change the strength and stability , and with it , the direction and stability of motivation. To check our results, we planned on using the battery of psychological techniques : 1. Schwartz technique that measures personal values and behavior in society, DORS procedure ( measurement of fatigue , monotony , satiety and stress ) and riffs that give the test once a week. Our assumption is

  19. Ruratae: A physics-based audio engine

    NASA Astrophysics Data System (ADS)

    Allen, Andrew S.

    In this thesis, I will demonstrate the capabilities of Ruratae, a physics-based audio rendering engine that models and sonifies mechanical vibrations of Newtonian bodies. This new system will allow its users a range of possibilities and subtle controls without requiring expert knowledge of signal theory and acoustics. It will be an environment that allows users to produce dynamic, reconfigurable, and interactive sounds through physically-intuitive construction and playing behaviors. I will discuss at detail known problems and issues that arise when attempting to afford these abilities, and I will offer several solutions and strategies that can be employed to tackle and reduce these upsets.

  20. Quantization and psychoacoustic model in audio coding in advanced audio coding

    NASA Astrophysics Data System (ADS)

    Brzuchalski, Grzegorz

    2011-10-01

    This paper presents complete optimized architecture of Advanced Audio Coder quantization with Huffman coding. After that psychoacoustic model theory is presented and few algorithms described: standard Two Loop Search, its modifications, Genetic, Just Noticeable Level Difference, Trellis-Based and its modification: Cascaded Trellis-Based Algorithm.

  1. Instructional Audio Guidelines: Four Design Principles to Consider for Every Instructional Audio Design Effort

    ERIC Educational Resources Information Center

    Carter, Curtis W.

    2012-01-01

    This article contends that instructional designers and developers should attend to four particular design principles when creating instructional audio. Support for this view is presented by referencing the limited research that has been done in this area, and by indicating how and why each of the four principles is important to the design process.…

  2. Investigating the impact of audio instruction and audio-visual biofeedback for lung cancer radiation therapy

    NASA Astrophysics Data System (ADS)

    George, Rohini

    Lung cancer accounts for 13% of all cancers in the Unites States and is the leading cause of deaths among both men and women. The five-year survival for lung cancer patients is approximately 15%.(ACS facts & figures) Respiratory motion decreases accuracy of thoracic radiotherapy during imaging and delivery. To account for respiration, generally margins are added during radiation treatment planning, which may cause a substantial dose delivery to normal tissues and increase the normal tissue toxicity. To alleviate the above-mentioned effects of respiratory motion, several motion management techniques are available which can reduce the doses to normal tissues, thereby reducing treatment toxicity and allowing dose escalation to the tumor. This may increase the survival probability of patients who have lung cancer and are receiving radiation therapy. However the accuracy of these motion management techniques are inhibited by respiration irregularity. The rationale of this thesis was to study the improvement in regularity of respiratory motion by breathing coaching for lung cancer patients using audio instructions and audio-visual biofeedback. A total of 331 patient respiratory motion traces, each four minutes in length, were collected from 24 lung cancer patients enrolled in an IRB-approved breathing-training protocol. It was determined that audio-visual biofeedback significantly improved the regularity of respiratory motion compared to free breathing and audio instruction, thus improving the accuracy of respiratory gated radiotherapy. It was also observed that duty cycles below 30% showed insignificant reduction in residual motion while above 50% there was a sharp increase in residual motion. The reproducibility of exhale based gating was higher than that of inhale base gating. Modeling the respiratory cycles it was found that cosine and cosine 4 models had the best correlation with individual respiratory cycles. The overall respiratory motion probability distribution

  3. 47 CFR 25.144 - Licensing provisions for the 2.3 GHz satellite digital audio radio service.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 47 Telecommunication 2 2010-10-01 2010-10-01 false Licensing provisions for the 2.3 GHz satellite digital audio radio service. 25.144 Section 25.144 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) COMMON CARRIER SERVICES SATELLITE COMMUNICATIONS Applications and Licenses Space Stations § 25.144 Licensing provisions for the 2.3...

  4. 47 CFR 25.144 - Licensing provisions for the 2.3 GHz satellite digital audio radio service.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 47 Telecommunication 2 2013-10-01 2013-10-01 false Licensing provisions for the 2.3 GHz satellite... (CONTINUED) COMMON CARRIER SERVICES SATELLITE COMMUNICATIONS Applications and Licenses Space Stations § 25.144 Licensing provisions for the 2.3 GHz satellite digital audio radio service. (a)...

  5. 47 CFR 25.144 - Licensing provisions for the 2.3 GHz satellite digital audio radio service.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 47 Telecommunication 2 2012-10-01 2012-10-01 false Licensing provisions for the 2.3 GHz satellite... (CONTINUED) COMMON CARRIER SERVICES SATELLITE COMMUNICATIONS Applications and Licenses Space Stations § 25.144 Licensing provisions for the 2.3 GHz satellite digital audio radio service. (a)...

  6. 47 CFR 25.144 - Licensing provisions for the 2.3 GHz satellite digital audio radio service.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 47 Telecommunication 2 2014-10-01 2014-10-01 false Licensing provisions for the 2.3 GHz satellite... (CONTINUED) COMMON CARRIER SERVICES SATELLITE COMMUNICATIONS Applications and Licenses Space Stations § 25.144 Licensing provisions for the 2.3 GHz satellite digital audio radio service. (a)...

  7. A Content-Adaptive Analysis and Representation Framework for Audio Event Discovery from "Unscripted" Multimedia

    NASA Astrophysics Data System (ADS)

    Radhakrishnan, Regunathan; Divakaran, Ajay; Xiong, Ziyou; Otsuka, Isao

    2006-12-01

    We propose a content-adaptive analysis and representation framework to discover events using audio features from "unscripted" multimedia such as sports and surveillance for summarization. The proposed analysis framework performs an inlier/outlier-based temporal segmentation of the content. It is motivated by the observation that "interesting" events in unscripted multimedia occur sparsely in a background of usual or "uninteresting" events. We treat the sequence of low/mid-level features extracted from the audio as a time series and identify subsequences that are outliers. The outlier detection is based on eigenvector analysis of the affinity matrix constructed from statistical models estimated from the subsequences of the time series. We define the confidence measure on each of the detected outliers as the probability that it is an outlier. Then, we establish a relationship between the parameters of the proposed framework and the confidence measure. Furthermore, we use the confidence measure to rank the detected outliers in terms of their departures from the background process. Our experimental results with sequences of low- and mid-level audio features extracted from sports video show that "highlight" events can be extracted effectively as outliers from a background process using the proposed framework. We proceed to show the effectiveness of the proposed framework in bringing out suspicious events from surveillance videos without any a priori knowledge. We show that such temporal segmentation into background and outliers, along with the ranking based on the departure from the background, can be used to generate content summaries of any desired length. Finally, we also show that the proposed framework can be used to systematically select "key audio classes" that are indicative of events of interest in the chosen domain.

  8. Availability of feature-oriented scanning probe microscopy for remote-controlled measurements on board a space laboratory or planet exploration Rover.

    PubMed

    Lapshin, Rostislav V

    2009-06-01

    Prospects for a feature-oriented scanning (FOS) approach to investigations of sample surfaces, at the micrometer and nanometer scales, with the use of scanning probe microscopy under space laboratory or planet exploration rover conditions, are examined. The problems discussed include decreasing sensitivity of the onboard scanning probe microscope (SPM) to temperature variations, providing autonomous operation, implementing the capabilities for remote control, self-checking, self-adjustment, and self-calibration. A number of topical problems of SPM measurements in outer space or on board a planet exploration rover may be solved via the application of recently proposed FOS methods. PMID:19566423

  9. Audio Utilization Conventions and Techniques for Computer Assisted Instruction.

    ERIC Educational Resources Information Center

    Army Signal Center and School, Fort Monmouth, NJ.

    A set of guidelines has been developed for the implementation of the audio mode in computer assisted instruction (CAI). The manual contains a collection of conventions and techniques synthesized from recent publications in areas pertinent to multi-media audiovisual presentation. These areas include audio message placement, positioning, frequency,…

  10. 47 CFR 10.520 - Common audio attention signal.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 47 Telecommunication 1 2011-10-01 2011-10-01 false Common audio attention signal. 10.520 Section 10.520 Telecommunication FEDERAL COMMUNICATIONS COMMISSION GENERAL COMMERCIAL MOBILE ALERT SYSTEM Equipment Requirements § 10.520 Common audio attention signal. A Participating CMS Provider and equipment manufacturers may only market devices...

  11. The Practical Audio-Visual Handbook for Teachers.

    ERIC Educational Resources Information Center

    Scuorzo, Herbert E.

    The use of audio/visual media as an aid to instruction is a common practice in today's classroom. Most teachers, however, have little or no formal training in this field and rarely a knowledgeable coordinator to help them. "The Practical Audio-Visual Handbook for Teachers" discusses the types and mechanics of many of these media forms and proposes…

  12. Audio and Text Density in Computer-based Instruction.

    ERIC Educational Resources Information Center

    Koroghlanian, Carol M.; Sullivan, Howard J.

    2000-01-01

    Investigates the effects of audio and text density on the achievement, time-in-program, and attitudes of 134 undergraduate students. Data concerning pre-existing computer skills/experience and demographic information were also collected. Findings have implications for the use of audio, text and graphics in the design of computer-based instruction.…

  13. Tune in the Net with RealAudio.

    ERIC Educational Resources Information Center

    Buchanan, Larry

    1997-01-01

    Describes how to connect to the RealAudio Web site to download a player that provides sound from Web pages to the computer through streaming technology. Explains hardware and software requirements and provides addresses for other RealAudio Web sites are provided, including weather information and current news. (LRW)

  14. Beyond Podcasting: Creative Approaches to Designing Educational Audio

    ERIC Educational Resources Information Center

    Middleton, Andrew

    2009-01-01

    This paper discusses a university-wide pilot designed to encourage academics to creatively explore learner-centred applications for digital audio. Participation in the pilot was diverse in terms of technical competence, confidence and contextual requirements and there was little prior experience of working with digital audio. Many innovative…

  15. Use of Audio Modification in Science Vocabulary Assessment

    ERIC Educational Resources Information Center

    Adiguzel, Tufan

    2011-01-01

    The purposes of this study were to examine the utilization of audio modification in vocabulary assessment in school subject areas, specifically in elementary science, and to present a web-based key vocabulary assessment tool for the elementary school level. Audio-recorded readings were used to replace independent student readings as the task…

  16. Use of Video and Audio Texts in EFL Listening Test

    ERIC Educational Resources Information Center

    Basal, Ahmet; Gülözer, Kaine; Demir, Ibrahim

    2015-01-01

    The study aims to discover whether audio or video modality in a listening test is more beneficial to test takers. In this study, the posttest-only control group design was utilized and quantitative data were collected in order to measure participant performances concerning two types of modality (audio or video) in a listening test. The…

  17. Effect of Audio vs. Video on Aural Discrimination of Vowels

    ERIC Educational Resources Information Center

    McCrocklin, Shannon

    2012-01-01

    Despite the growing use of media in the classroom, the effects of using of audio versus video in pronunciation teaching has been largely ignored. To analyze the impact of the use of audio or video training on aural discrimination of vowels, 61 participants (all students at a large American university) took a pre-test followed by two training…

  18. A Case Study on Audio Feedback with Geography Undergraduates

    ERIC Educational Resources Information Center

    Rodway-Dyer, Sue; Knight, Jasper; Dunne, Elizabeth

    2011-01-01

    Several small-scale studies have suggested that audio feedback can help students to reflect on their learning and to develop deep learning approaches that are associated with higher attainment in assessments. For this case study, Geography undergraduates were given audio feedback on a written essay assignment, alongside traditional written…

  19. Using Audio Books to Improve Reading and Academic Performance

    ERIC Educational Resources Information Center

    Montgomery, Joel R.

    2009-01-01

    This article highlights significant research about what below grade-level reading means in middle school classrooms and suggests a tested approach to improve reading comprehension levels significantly by using audio books. The use of these audio books can improve reading and academic performance for both English language learners (ELLs) and for…

  20. An Audio Stream Redirector for the Ethernet Speaker

    ERIC Educational Resources Information Center

    Mandrekar, Ishan; Prevelakis, Vassilis; Turner, David Michael

    2004-01-01

    The authors have developed the "Ethernet Speaker" (ES), a network-enabled single board computer embedded into a conventional audio speaker. Audio streams are transmitted in the local area network using multicast packets, and the ES can select any one of them and play it back. A key requirement for the ES is that it must be capable of playing any…

  1. The Audio-Visual Equipment Directory. Seventeenth Edition.

    ERIC Educational Resources Information Center

    Herickes, Sally, Ed.

    The following types of audiovisual equipment are catalogued: 8 mm. and 16 mm. motion picture projectors, filmstrip and sound filmstrip projectors, slide projectors, random access projection equipment, opaque, overhead, and micro-projectors, record players, special purpose projection equipment, audio tape recorders and players, audio tape…

  2. Making the Most of Audio. Technology in Language Learning Series.

    ERIC Educational Resources Information Center

    Barley, Anthony

    Prepared for practicing language teachers, this book's aim is to help them make the most of audio, a readily accessible resource. The book shows, with the help of numerous practical examples, how a range of language skills can be developed. Most examples are in French. Chapters cover the following information: (1) making the most of audio (e.g.,…

  3. Audio Podcasting in a Tablet PC-Enhanced Biochemistry Course

    ERIC Educational Resources Information Center

    Lyles, Heather; Robertson, Brian; Mangino, Michael; Cox, James R.

    2007-01-01

    This report describes the effects of making audio podcasts of all lectures in a large, basic biochemistry course promptly available to students. The audio podcasts complement a previously described approach in which a tablet PC is used to annotate PowerPoint slides with digital ink to produce electronic notes that can be archived. The fundamentals…

  4. Effective Use of Audio Media in Multimedia Presentations.

    ERIC Educational Resources Information Center

    Kerr, Brenda

    This paper emphasizes research-based reasons for adding audio to multimedia presentations. The first section summarizes suggestions from a review of research on the effectiveness of audio media when accompanied by other forms of media; types of research studies (e.g., evaluation, intra-medium, and aptitude treatment interaction studies) are also…

  5. 43 CFR 8365.2-2 - Audio devices.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 43 Public Lands: Interior 2 2013-10-01 2013-10-01 false Audio devices. 8365.2-2 Section 8365.2-2 Public Lands: Interior Regulations Relating to Public Lands (Continued) BUREAU OF LAND MANAGEMENT, DEPARTMENT OF THE INTERIOR RECREATION PROGRAMS VISITOR SERVICES Rules of Conduct § 8365.2-2 Audio devices....

  6. An Exploration of the Dust Spectral Features of the Carbon-Rich Star V Cyg Through Time and Space

    NASA Astrophysics Data System (ADS)

    Reel, Matthew; Speck, A.; Sloan, G. C.; Volk, K.

    2014-01-01

    Carbon-rich AGB stars are surrounded by circumstellar shells of gas and dust. The dust is dominated by carbon (probably graphitic) and silicon carbide (SiC), which is the source of the observed 11μm spectral feature. We investigate the nearby carbon star V Cyg which has been observed numerous times over the past few decades. By analyzing the temporal spectral variations associated with the stellar pulsation cycle we study how the pulsation cycle affects the circumstellar dust. The Infrared spectrum of the star also shows many prominent molecular absorption bands, as well as the “30μm” emission feature which has previously been attributed to magnesium sulfide (MgS). In addition to the temporal data, we have also obtained spatially-resolved spectroscopic data for the dust shell(s) around V Cyg. Combining these various spectral observations with AAVSO data on variations in the visual magnitude we investigate temporal variations in both the 11μm and 30μm spectral features and sought to correlate these temporal variations with the spatial variations in the 11μm feature. Our results indicate many spatial correlations within the 11μm feature parameters, as well as correlations which may suggest a carrier of the 30μm feature other than pure MgS.

  7. TVAR modeling of EEG to detect audio distraction during simulated driving

    NASA Astrophysics Data System (ADS)

    Dahal, Nabaraj; (Nanda Nandagopal, D.; Cocks, Bernadine; Vijayalakshmi, Ramasamy; Dasari, Naga; Gaertner, Paul

    2014-06-01

    Objective. The objective of our current study was to look for the EEG correlates that can reveal the engaged state of the brain while undertaking cognitive tasks. Specifically, we aimed to identify EEG features that could detect audio distraction during simulated driving. Approach. Time varying autoregressive (TVAR) analysis using Kalman smoother was carried out on short time epochs of EEG data collected from participants as they undertook two simulated driving tasks. TVAR coefficients were then used to construct all pole model enabling the identification of EEG features that could differentiate normal driving from audio distracted driving. Main results. Pole analysis of the TVAR model led to the visualization of event related synchronization/desynchronization (ERS/ERD) patterns in the form of pole displacements in pole plots of the temporal EEG channels in the z plane enabling the differentiation of the two driving conditions. ERS in the EEG data has been demonstrated during audio distraction as an associated phenomenon. Significance. Visualizing the ERD/ERS phenomenon in terms of pole displacement is a novel approach. Although ERS/ERD has previously been demonstrated as reliable when applied to motor related tasks, it is believed to be the first time that it has been applied to investigate human cognitive phenomena such as attention and distraction. Results confirmed that distracted/non-distracted driving states can be identified using this approach supporting its applicability to cognition research.

  8. Effect of downsampling and compressive sensing on audio-based continuous cough monitoring.

    PubMed

    Casaseca-de-la-Higuera, Pablo; Lesso, Paul; McKinstry, Brian; Pinnock, Hilary; Rabinovich, Roberto; McCloughan, Lucy; Monge-Álvarez, Jesús

    2015-01-01

    This paper presents an efficient cough detection system based on simple decision-tree classification of spectral features from a smartphone audio signal. Preliminary evaluation on voluntary coughs shows that the system can achieve 98% sensitivity and 97.13% specificity when the audio signal is sampled at full rate. With this baseline system, we study possible efficiency optimisations by evaluating the effect of downsampling below the Nyquist rate and how the system performance at low sampling frequencies can be improved by incorporating compressive sensing reconstruction schemes. Our results show that undersampling down to 400 Hz can still keep sensitivity and specificity values above 90% despite of aliasing. Furthermore, the sparsity of cough signals in the time domain allows keeping performance figures close to 90% when sampling at 100 Hz using compressive sensing schemes. PMID:26737716

  9. Transitioning from analog to digital audio recording in childhood speech sound disorders

    PubMed Central

    Shriberg, Lawrence D.; McSweeny, Jane L.; Anderson, Bruce E.; Campbell, Thomas F.; Chial, Michael R.; Green, Jordan R.; Hauner, Katherina K.; Moore, Christopher A.; Rusiewicz, Heather L.; Wilson, David L.

    2014-01-01

    Few empirical findings or technical guidelines are available on the current transition from analog to digital audio recording in childhood speech sound disorders. Of particular concern in the present context was whether a transition from analog- to digital-based transcription and coding of prosody and voice features might require re-standardizing a reference database for research in childhood speech sound disorders. Two research transcribers with different levels of experience glossed, transcribed, and prosody-voice coded conversational speech samples from eight children with mild to severe speech disorders of unknown origin. The samples were recorded, stored, and played back using representative analog and digital audio systems. Effect sizes calculated for an array of analog versus digital comparisons ranged from negligible to medium, with a trend for participants’ speech competency scores to be slightly lower for samples obtained and transcribed using the digital system. We discuss the implications of these and other findings for research and clinical practise. PMID:16019779

  10. Audio Haptic Videogaming for Developing Wayfinding Skills in Learners Who are Blind.

    PubMed

    Sánchez, Jaime; de Borba Campos, Marcia; Espinoza, Matías; Merabet, Lotfi B

    2014-01-01

    Interactive digital technologies are currently being developed as a novel tool for education and skill development. Audiopolis is an audio and haptic based videogame designed for developing orientation and mobility (O&M) skills in people who are blind. We have evaluated the cognitive impact of videogame play on O&M skills by assessing performance on a series of behavioral tasks carried out in both indoor and outdoor virtual spaces. Our results demonstrate that the use of Audiopolis had a positive impact on the development and use of O&M skills in school-aged learners who are blind. The impact of audio and haptic information on learning is also discussed. PMID:25485312

  11. Audio Haptic Videogaming for Developing Wayfinding Skills in Learners Who are Blind

    PubMed Central

    Sánchez, Jaime; de Borba Campos, Marcia; Espinoza, Matías; Merabet, Lotfi B.

    2014-01-01

    Interactive digital technologies are currently being developed as a novel tool for education and skill development. Audiopolis is an audio and haptic based videogame designed for developing orientation and mobility (O&M) skills in people who are blind. We have evaluated the cognitive impact of videogame play on O&M skills by assessing performance on a series of behavioral tasks carried out in both indoor and outdoor virtual spaces. Our results demonstrate that the use of Audiopolis had a positive impact on the development and use of O&M skills in school-aged learners who are blind. The impact of audio and haptic information on learning is also discussed. PMID:25485312

  12. ESA personal communications and digital audio broadcasting systems based on non-geostationary satellites

    NASA Technical Reports Server (NTRS)

    Logalbo, P.; Benedicto, J.; Viola, R.

    1993-01-01

    Personal Communications and Digital Audio Broadcasting are two new services that the European Space Agency (ESA) is investigating for future European and Global Mobile Satellite systems. ESA is active in promoting these services in their various mission options including non-geostationary and geostationary satellite systems. A Medium Altitude Global Satellite System (MAGSS) for global personal communications at L and S-band, and a Multiregional Highly inclined Elliptical Orbit (M-HEO) system for multiregional digital audio broadcasting at L-band are described. Both systems are being investigated by ESA in the context of future programs, such as Archimedes, which are intended to demonstrate the new services and to develop the technology for future non-geostationary mobile communication and broadcasting satellites.

  13. Audio/Visual Aids: A Study of the Effect of Audio/Visual Aids on the Comprehension Recall of Students.

    ERIC Educational Resources Information Center

    Bavaro, Sandra

    A study investigated whether the use of audio/visual aids had an effect upon comprehension recall. Thirty fourth-grade students from an urban public school were randomly divided into two equal samples of 15. One group was given a story to read (print only), while the other group viewed a filmstrip of the same story, thereby utilizing audio/visual…

  14. Horatio Audio-Describes Shakespeare's "Hamlet": Blind and Low-Vision Theatre-Goers Evaluate an Unconventional Audio Description Strategy

    ERIC Educational Resources Information Center

    Udo, J. P.; Acevedo, B.; Fels, D. I.

    2010-01-01

    Audio description (AD) has been introduced as one solution for providing people who are blind or have low vision with access to live theatre, film and television content. However, there is little research to inform the process, user preferences and presentation style. We present a study of a single live audio-described performance of Hart House…

  15. On Building Immersive Audio Applications Using Robust Adaptive Beamforming and Joint Audio-Video Source Localization

    NASA Astrophysics Data System (ADS)

    Beracoechea, J. A.; Torres-Guijarro, S.; García, L.; Casajús-Quirós, F. J.

    2006-12-01

    This paper deals with some of the different problems, strategies, and solutions of building true immersive audio systems oriented to future communication applications. The aim is to build a system where the acoustic field of a chamber is recorded using a microphone array and then is reconstructed or rendered again, in a different chamber using loudspeaker array-based techniques. Our proposal explores the possibility of using recent robust adaptive beamforming techniques for effectively estimating the original sources of the emitting room. A joint audio-video localization method needed in the estimation process as well as in the rendering engine is also presented. The estimated source signal and the source localization information drive a wave field synthesis engine that renders the acoustic field again at the receiving chamber. The system performance is tested using MUSHRA-based subjective tests.

  16. Frequency dependent squeezed light at audio frequencies

    NASA Astrophysics Data System (ADS)

    Miller, John

    2015-04-01

    Following successful implementation in the previous generation of instruments, squeezed states of light represent a proven technology for the reduction of quantum noise in ground-based interferometric gravitational-wave detectors. As a result of lower noise and increased circulating power, the current generation of detectors places one further demand on this technique - that the orientation of the squeezed ellipse be rotated as function of frequency. This extension allows previously negligible quantum radiation pressure noise to be mitigated in addition to quantum shot noise. I will present the results of an experiment which performs the appropriate rotation by reflecting the squeezed state from a detuned high-finesse optical cavity, demonstrating frequency dependent squeezing at audio frequencies for the first time and paving the way for broadband quantum noise reduction in Advanced LIGO. Further, I will indicate how a realistic implementation of this approach will impact Advanced LIGO both alone and in combination with other potential upgrades.

  17. Spatial audio through a bone conduction interface.

    PubMed

    MacDonald, Justin A; Henry, Paula P; Letowski, Tomasz R

    2006-10-01

    Headphones are the standard presentation device for radio communication in the military. Although bone conduction devices possess several advantages over headphones for some military applications, they are generally considered inappropriate for inclusion in a multi-channel system. The current study tested the feasibility of a multi-channel bone conduction system by measuring the localizability of spatialized auditory stimuli presented through a pair of bone conduction vibrators. Listeners localized a Gaussian noise stimulus spatialized with individualized head-related transfer functions (HRTFs). The sounds were presented from eight virtual locations on the horizontal plane (0, +/-45, +/-90, +/-135, and 180 degrees ) through either stereo headphones or a stereo bone conduction system. Localization performance was found to be nearly identical for both audio systems, indicating that bone conduction systems can be effectively used for displaying spatial information. PMID:17062501

  18. A direct broadcast satellite-audio experiment

    NASA Technical Reports Server (NTRS)

    Vaisnys, Arvydas; Abbe, Brian; Motamedi, Masoud

    1992-01-01

    System studies have been carried out over the past three years at the Jet Propulsion Laboratory (JPL) on digital audio broadcasting (DAB) via satellite. The thrust of the work to date has been on designing power and bandwidth efficient systems capable of providing reliable service to fixed, mobile, and portable radios. It is very difficult to predict performance in an environment which produces random periods of signal blockage, such as encountered in mobile reception where a vehicle can quickly move from one type of terrain to another. For this reason, some signal blockage mitigation techniques were built into an experimental DAB system and a satellite experiment was conducted to obtain both qualitative and quantitative measures of performance in a range of reception environments. This paper presents results from the experiment and some conclusions on the effectiveness of these blockage mitigation techniques.

  19. Thermal and neutron-physical features of the nuclear reactor for a power pulsation plant for space applications

    NASA Astrophysics Data System (ADS)

    Gordeev, É. G.; Kaminskii, A. S.; Konyukhov, G. V.; Pavshuk, V. A.; Turbina, T. A.

    2012-05-01

    We have explored the possibility of creating small-size reactors with a high power output with the provision of thermal stability and nuclear safety under standard operating conditions and in emergency situations. The neutron-physical features of such a reactor have been considered and variants of its designs preserving the main principles and approaches of nuclear rocket engine technology are presented.

  20. Materials Science Research Hardware for Application on the International Space Station: an Overview of Typical Hardware Requirements and Features

    NASA Technical Reports Server (NTRS)

    Schaefer, D. A.; Cobb, S.; Fiske, M. R.; Srinivas, R.

    2000-01-01

    NASA's Marshall Space Flight Center (MSFC) is the lead center for Materials Science Microgravity Research. The Materials Science Research Facility (MSRF) is a key development effort underway at MSFC. The MSRF will be the primary facility for microgravity materials science research on board the International Space Station (ISS) and will implement the NASA Materials Science Microgravity Research Program. It will operate in the U.S. Laboratory Module and support U. S. Microgravity Materials Science Investigations. This facility is being designed to maintain the momentum of the U.S. role in microgravity materials science and support NASA's Human Exploration and Development of Space (HEDS) Enterprise goals and objectives for Materials Science. The MSRF as currently envisioned will consist of three Materials Science Research Racks (MSRR), which will be deployed to the International Space Station (ISS) in phases, Each rack is being designed to accommodate various Experiment Modules, which comprise processing facilities for peer selected Materials Science experiments. Phased deployment will enable early opportunities for the U.S. and International Partners, and support the timely incorporation of technology updates to the Experiment Modules and sensor devices.

  1. Effects of aging on audio-visual speech integration.

    PubMed

    Huyse, Aurélie; Leybaert, Jacqueline; Berthommier, Frédéric

    2014-10-01

    This study investigated the impact of aging on audio-visual speech integration. A syllable identification task was presented in auditory-only, visual-only, and audio-visual congruent and incongruent conditions. Visual cues were either degraded or unmodified. Stimuli were embedded in stationary noise alternating with modulated noise. Fifteen young adults and 15 older adults participated in this study. Results showed that older adults had preserved lipreading abilities when the visual input was clear but not when it was degraded. The impact of aging on audio-visual integration also depended on the quality of the visual cues. In the visual clear condition, the audio-visual gain was similar in both groups and analyses in the framework of the fuzzy-logical model of perception confirmed that older adults did not differ from younger adults in their audio-visual integration abilities. In the visual reduction condition, the audio-visual gain was reduced in the older group, but only when the noise was stationary, suggesting that older participants could compensate for the loss of lipreading abilities by using the auditory information available in the valleys of the noise. The fuzzy-logical model of perception confirmed the significant impact of aging on audio-visual integration by showing an increased weight of audition in the older group. PMID:25324091

  2. An inconclusive digital audio authenticity examination: a unique case.

    PubMed

    Koenig, Bruce E; Lacey, Douglas S

    2012-01-01

    This case report sets forth an authenticity examination of 35 encrypted, proprietary-format digital audio files containing recorded telephone conversations between two codefendants in a criminal matter. The codefendant who recorded the conversations did so on a recording system he developed; additionally, he was both a forensic audio authenticity examiner, who had published and presented in the field, and was the head of a professional audio society's writing group for authenticity standards. The authors conducted the examination of the recordings following nine laboratory steps of the peer-reviewed and published 11-step digital audio authenticity protocol. Based considerably on the codefendant's direct involvement with the development of the encrypted audio format, his experience in the field of forensic audio authenticity analysis, and the ease with which the audio files could be accessed, converted, edited in the gap areas, and reconstructed in such a way that the processes were undetected, the authors concluded that the recordings could not be scientifically authenticated through accepted forensic practices. PMID:21854384

  3. Musical examination to bridge audio data and sheet music

    NASA Astrophysics Data System (ADS)

    Pan, Xunyu; Cross, Timothy J.; Xiao, Liangliang; Hei, Xiali

    2015-03-01

    The digitalization of audio is commonly implemented for the purpose of convenient storage and transmission of music and songs in today's digital age. Analyzing digital audio for an insightful look at a specific musical characteristic, however, can be quite challenging for various types of applications. Many existing musical analysis techniques can examine a particular piece of audio data. For example, the frequency of digital sound can be easily read and identified at a specific section in an audio file. Based on this information, we could determine the musical note being played at that instant, but what if you want to see a list of all the notes played in a song? While most existing methods help to provide information about a single piece of the audio data at a time, few of them can analyze the available audio file on a larger scale. The research conducted in this work considers how to further utilize the examination of audio data by storing more information from the original audio file. In practice, we develop a novel musical analysis system Musicians Aid to process musical representation and examination of audio data. Musicians Aid solves the previous problem by storing and analyzing the audio information as it reads it rather than tossing it aside. The system can provide professional musicians with an insightful look at the music they created and advance their understanding of their work. Amateur musicians could also benefit from using it solely for the purpose of obtaining feedback about a song they were attempting to play. By comparing our system's interpretation of traditional sheet music with their own playing, a musician could ensure what they played was correct. More specifically, the system could show them exactly where they went wrong and how to adjust their mistakes. In addition, the application could be extended over the Internet to allow users to play music with one another and then review the audio data they produced. This would be particularly

  4. Video genre classification using multimodal features

    NASA Astrophysics Data System (ADS)

    Jin, Sung Ho; Bae, Tae Meon; Choo, Jin Ho; Ro, Yong Man

    2003-12-01

    We propose a video genre classification method using multimodal features. The proposed method is applied for the preprocessing of automatic video summarization or the retrieval and classification of broadcasting video contents. Through a statistical analysis of low-level and middle-level audio-visual features in video, the proposed method can achieve good performance in classifying several broadcasting genres such as cartoon, drama, music video, news, and sports. In this paper, we adopt MPEG-7 audio-visual descriptors as multimodal features of video contents and evaluate the performance of the classification by feeding the features into a decision tree-based classifier which is trained by CART. The experimental results show that the proposed method can recognize several broadcasting video genres with a high accuracy and the classification performance with multimodal features is superior to the one with unimodal features in the genre classification.

  5. Towards Structural Analysis of Audio Recordings in the Presence of Musical Variations

    NASA Astrophysics Data System (ADS)

    Müller, Meinard; Kurth, Frank

    2006-12-01

    One major goal of structural analysis of an audio recording is to automatically extract the repetitive structure or, more generally, the musical form of the underlying piece of music. Recent approaches to this problem work well for music, where the repetitions largely agree with respect to instrumentation and tempo, as is typically the case for popular music. For other classes of music such as Western classical music, however, musically similar audio segments may exhibit significant variations in parameters such as dynamics, timbre, execution of note groups, modulation, articulation, and tempo progression. In this paper, we propose a robust and efficient algorithm for audio structure analysis, which allows to identify musically similar segments even in the presence of large variations in these parameters. To account for such variations, our main idea is to incorporate invariance at various levels simultaneously: we design a new type of statistical features to absorb microvariations, introduce an enhanced local distance measure to account for local variations, and describe a new strategy for structure extraction that can cope with the global variations. Our experimental results with classical and popular music show that our algorithm performs successfully even in the presence of significant musical variations.

  6. The effect of reverberation on personal audio devices.

    PubMed

    Simón-Gálvez, Marcos F; Elliott, Stephen J; Cheer, Jordan

    2014-05-01

    Personal audio refers to the creation of a listening zone within which a person, or a group of people, hears a given sound program, without being annoyed by other sound programs being reproduced in the same space. Generally, these different sound zones are created by arrays of loudspeakers. Although these devices have the capacity to achieve different sound zones in an anechoic environment, they are ultimately used in normal rooms, which are reverberant environments. At high frequencies, reflections from the room surfaces create a diffuse pressure component which is uniform throughout the room volume and thus decreases the directional characteristics of the device. This paper shows how the reverberant performance of an array can be modeled, knowing the anechoic performance of the radiator and the acoustic characteristics of the room. A formulation is presented whose results are compared to practical measurements in reverberant environments. Due to reflections from the room surfaces, pressure variations are introduced in the transfer responses of the array. This aspect is assessed by means of simulations where random noise is added to create uncertainties, and by performing measurements in a real environment. These results show how the robustness of an array is increased when it is designed for use in a reverberant environment. PMID:24815249

  7. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 17 Commodity and Securities Exchanges 2 2010-04-01 2010-04-01 false Graphic, image, audio and... Submissions § 232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio..., image, audio or video material is presented in the delivered version, or they may be listed in...

  8. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... 17 Commodity and Securities Exchanges 2 2011-04-01 2011-04-01 false Graphic, image, audio and... Submissions § 232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio..., image, audio or video material is presented in the delivered version, or they may be listed in...

  9. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2014 CFR

    2014-04-01

    ... 17 Commodity and Securities Exchanges 3 2014-04-01 2014-04-01 false Graphic, image, audio and... Submissions § 232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio..., image, audio or video material is presented in the delivered version, or they may be listed in...

  10. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2013 CFR

    2013-04-01

    ... 17 Commodity and Securities Exchanges 2 2013-04-01 2013-04-01 false Graphic, image, audio and... Submissions § 232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio..., image, audio or video material is presented in the delivered version, or they may be listed in...

  11. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2012 CFR

    2012-04-01

    ... 17 Commodity and Securities Exchanges 2 2012-04-01 2012-04-01 false Graphic, image, audio and... Submissions § 232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio..., image, audio or video material is presented in the delivered version, or they may be listed in...

  12. 37 CFR 201.27 - Initial notice of distribution of digital audio recording devices or media.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... distribution of digital audio recording devices or media. 201.27 Section 201.27 Patents, Trademarks, and... Initial notice of distribution of digital audio recording devices or media. (a) General. This section..., any digital audio recording device or digital audio recording medium in the United States....

  13. 37 CFR 201.28 - Statements of Account for digital audio recording devices or media.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... digital audio recording devices or media. 201.28 Section 201.28 Patents, Trademarks, and Copyrights... of Account for digital audio recording devices or media. (a) General. This section prescribes rules... United States any digital audio recording device or digital audio recording medium. (b) Definitions....

  14. 37 CFR 201.28 - Statements of Account for digital audio recording devices or media.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... digital audio recording devices or media. 201.28 Section 201.28 Patents, Trademarks, and Copyrights... of Account for digital audio recording devices or media. (a) General. This section prescribes rules... United States any digital audio recording device or digital audio recording medium. (b) Definitions....

  15. 37 CFR 201.27 - Initial notice of distribution of digital audio recording devices or media.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... distribution of digital audio recording devices or media. 201.27 Section 201.27 Patents, Trademarks, and... Initial notice of distribution of digital audio recording devices or media. (a) General. This section..., any digital audio recording device or digital audio recording medium in the United States....

  16. 37 CFR 201.27 - Initial notice of distribution of digital audio recording devices or media.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... distribution of digital audio recording devices or media. 201.27 Section 201.27 Patents, Trademarks, and... Initial notice of distribution of digital audio recording devices or media. (a) General. This section..., any digital audio recording device or digital audio recording medium in the United States....

  17. 37 CFR 201.27 - Initial notice of distribution of digital audio recording devices or media.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... distribution of digital audio recording devices or media. 201.27 Section 201.27 Patents, Trademarks, and... § 201.27 Initial notice of distribution of digital audio recording devices or media. (a) General. This... and distribute, any digital audio recording device or digital audio recording medium in the...

  18. 37 CFR 201.28 - Statements of Account for digital audio recording devices or media.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... digital audio recording devices or media. 201.28 Section 201.28 Patents, Trademarks, and Copyrights U.S... of Account for digital audio recording devices or media. (a) General. This section prescribes rules... United States any digital audio recording device or digital audio recording medium. (b) Definitions....

  19. 37 CFR 201.28 - Statements of Account for digital audio recording devices or media.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... digital audio recording devices or media. 201.28 Section 201.28 Patents, Trademarks, and Copyrights... of Account for digital audio recording devices or media. (a) General. This section prescribes rules... United States any digital audio recording device or digital audio recording medium. (b) Definitions....

  20. Performance Characterization of Loctite (Registered Trademark) 242 and 271 Liquid Locking Compounds (LLCs) as a Secondary Locking Feature for International Space Station (ISS) Fasteners

    NASA Technical Reports Server (NTRS)

    Dube, Michael J.; Gamwell, Wayne R.

    2011-01-01

    Several International Space Station (ISS) hardware components use Loctite (and other polymer based liquid locking compounds (LLCs)) as a means of meeting the secondary (redundant) locking feature requirement for fasteners. The primary locking method is the fastener preload, with the application of the Loctite compound which when cured is intended to resist preload reduction. The reliability of these compounds has been questioned due to a number of failures during ground testing. The ISS Program Manager requested the NASA Engineering and Safety Center (NESC) to characterize and quantify sensitivities of Loctite being used as a secondary locking feature. The findings and recommendations provided in this investigation apply to the anaerobic LLCs Loctite 242 and 271. No other anaerobic LLCs were evaluated for this investigation. This document contains the findings and recommendations of the NESC investigation

  1. Biomedical image representation approach using visualness and spatial information in a concept feature space for interactive region-of-interest-based retrieval.

    PubMed

    Rahman, Md Mahmudur; Antani, Sameer K; Demner-Fushman, Dina; Thoma, George R

    2015-10-01

    This article presents an approach to biomedical image retrieval by mapping image regions to local concepts where images are represented in a weighted entropy-based concept feature space. The term "concept" refers to perceptually distinguishable visual patches that are identified locally in image regions and can be mapped to a glossary of imaging terms. Further, the visual significance (e.g., visualness) of concepts is measured as the Shannon entropy of pixel values in image patches and is used to refine the feature vector. Moreover, the system can assist the user in interactively selecting a region-of-interest (ROI) and searching for similar image ROIs. Further, a spatial verification step is used as a postprocessing step to improve retrieval results based on location information. The hypothesis that such approaches would improve biomedical image retrieval is validated through experiments on two different data sets, which are collected from open access biomedical literature. PMID:26730398

  2. Nonlinear dimensionality reduction of CT histogram based feature space for predicting recurrence-free survival in non-small-cell lung cancer

    NASA Astrophysics Data System (ADS)

    Kawata, Y.; Niki, N.; Ohmatsu, H.; Aokage, K.; Kusumoto, M.; Tsuchida, T.; Eguchi, K.; Kaneko, M.

    2015-03-01

    Advantages of CT scanners with high resolution have allowed the improved detection of lung cancers. In the recent release of positive results from the National Lung Screening Trial (NLST) in the US showing that CT screening does in fact have a positive impact on the reduction of lung cancer related mortality. While this study does show the efficacy of CT based screening, physicians often face the problems of deciding appropriate management strategies for maximizing patient survival and for preserving lung function. Several key manifold-learning approaches efficiently reveal intrinsic low-dimensional structures latent in high-dimensional data spaces. This study was performed to investigate whether the dimensionality reduction can identify embedded structures from the CT histogram feature of non-small-cell lung cancer (NSCLC) space to improve the performance in predicting the likelihood of RFS for patients with NSCLC.

  3. How to Use Both Audio Channels on Video Cassette Recorders.

    ERIC Educational Resources Information Center

    Senour, Robert A.; Campbell, Lloyd

    1980-01-01

    Videotape cassettes have two audio channels which makes possible a wide variety of new instructional techniques. Instructors can put running commentary/critiques on tapes of student counselor training sessions. (Author/TG)

  4. Effectiveness of audio-based instruction in medical pharmacology.

    PubMed

    Bogner, P; Sajid, A W; Ford, D L

    1975-07-01

    An audio-based method of instruction on the topic of diuretic drugs was compared with the lecture method using a random sample of 66 sophomore medical students. A cognitive test given to measure achievement as a result of the two treatments showed that the average score of 80 percent for the audio group was significantly higher (p smaller than 0.25) than the 75 percent for the lecture group. A comparison of the lowest quartile scores in each group revealed that the audio-based group also scored higher 13 points, p smaller than .05) than the lecture group; however, no difference was noted in the upper three quartile scores of the two groups. The amount of time spent studying did not appear to be a factor. A second test given to meausre attitudes toward the audio method showed strong positive views among students. PMID:48549

  5. Proper Use of Audio-Visual Aids: Essential for Educators.

    ERIC Educational Resources Information Center

    Dejardin, Conrad

    1989-01-01

    Criticizes educators as the worst users of audio-visual aids and among the worst public speakers. Offers guidelines for the proper use of an overhead projector and the development of transparencies. (DMM)

  6. Audio CAPTCHA for SIP-Based VoIP

    NASA Astrophysics Data System (ADS)

    Soupionis, Yannis; Tountas, George; Gritzalis, Dimitris

    Voice over IP (VoIP) introduces new ways of communication, while utilizing existing data networks to provide inexpensive voice communications worldwide as a promising alternative to the traditional PSTN telephony. SPam over Internet Telephony (SPIT) is one potential source of future annoyance in VoIP. A common way to launch a SPIT attack is the use of an automated procedure (bot), which generates calls and produces audio advertisements. In this paper, our goal is to design appropriate CAPTCHA to fight such bots. We focus on and develop audio CAPTCHA, as the audio format is more suitable for VoIP environments and we implement it in a SIP-based VoIP environment. Furthermore, we suggest and evaluate the specific attributes that audio CAPTCHA should incorporate in order to be effective, and test it against an open source bot implementation.

  7. Robustness evaluation of transactional audio watermarking systems

    NASA Astrophysics Data System (ADS)

    Neubauer, Christian; Steinebach, Martin; Siebenhaar, Frank; Pickel, Joerg

    2003-06-01

    Distribution via Internet is of increasing importance. Easy access, transmission and consumption of digitally represented music is very attractive to the consumer but led also directly to an increasing problem of illegal copying. To cope with this problem watermarking is a promising concept since it provides a useful mechanism to track illicit copies by persistently attaching property rights information to the material. Especially for online music distribution the use of so-called transaction watermarking, also denoted with the term bitstream watermarking, is beneficial since it offers the opportunity to embed watermarks directly into perceptually encoded material without the need of full decompression/compression. Besides the concept of bitstream watermarking, former publications presented the complexity, the audio quality and the detection performance. These results are now extended by an assessment of the robustness of such schemes. The detection performance before and after applying selected attacks is presented for MPEG-1/2 Layer 3 (MP3) and MPEG-2/4 AAC bitstream watermarking, contrasted to the performance of PCM spread spectrum watermarking.

  8. Concepts, strategies and potentials using hypo-g and other features of the space environment for commercialization using higher plants

    NASA Technical Reports Server (NTRS)

    Krikorian, A. D.

    1985-01-01

    Opportunities for releasing, capturing, constructing and/or fixing the differential expressions or response potentials of the higher plant genome in the hypo-g environment for commercialization are explored. General strategies include improved plant-growing, crop and forestry production systems which conserve soil, water, labor and energy resources, and nutritional partitioning and mobilization of nutrients and synthates. Tissue and cell culture techniques of commercial potential include the growing and manipulation of cultured plant cells in vitro in a bioreactor to produce biologicals and secondary plants of economic value. The facilitation of plant breeding, the cloning of specific pathogen-free materials, the elimination of growing point or apex viruses, and the increase of plant yield are other O-g applications. The space environment may be advantageous in somatic embryogenesis, the culture of alkaloids, and the development of completely new crop plant germ plasm.

  9. Experiences with audio feedback in a veterinary curriculum.

    PubMed

    Rhind, Susan M; Pettigrew, Graham W; Spiller, Jo; Pearson, Geoff T

    2013-01-01

    On a national scale in the United Kingdom, student surveys have served to highlight areas within higher education that are not achieving high student satisfaction. Of particular concern to the veterinary and medical disciplines are the persistently poor levels of student satisfaction with academic feedback compared to students in other subjects. In this study we describe experiences with audio feedback trials in a veterinary curriculum. Students received audio feedback on either an in-course laboratory practical report or on an in-course multiple-choice test. Shortly after receiving their feedback, students were surveyed using an electronic questionnaire. In both courses, more students strongly agreed that audio feedback was helpful compared to either text-based (course A) or whole-class (course B) feedback. When asked to reflect on the helpfulness of various types of feedback they had received, audio feedback was rated less helpful than individual discussion with a member of staff (course A and course B), more helpful than peer discussion or automated feedback (course A and course B), and more helpful than written comments or whole-class review sessions (course B). From a faculty perspective, in course A, use of audio feedback was more efficient than handwritten feedback. In course B, the additional time commitment required was approximately 5 hours. Major themes in the qualitative data included the personal and individual nature of the feedback, quantity of feedback, improvement in students' insight into the process of marking, and the capacity of audio feedback to encourage and motivate. PMID:23470242

  10. The power of digital audio in interactive instruction: An unexploited medium

    SciTech Connect

    Pratt, J.; Trainor, M.

    1989-01-01

    Widespread use of audio in computer-based training (CBT) occurred with the advent of the interactive videodisc technology. This paper discusses the alternative of digital audio, which, unlike videodisc audio, enables one to rapidly revise the audio used in the CBT and which may be used in nonvideo CBT applications as well. We also discuss techniques used in audio script writing, editing, and production. Results from evaluations indicate a high degree of user satisfaction. 4 refs.

  11. Full-Featured Web Conferencing Systems

    ERIC Educational Resources Information Center

    Foreman, Joel; Jenkins, Roy

    2005-01-01

    In order to match the customary strengths of the still dominant face-to-face instructional mode, a high-performance online learning system must employ synchronous as well as asynchronous communications; buttress graphics, animation, and text with live audio and video; and provide many of the features and processes associated with course management…

  12. Digital Audio Radio Broadcast Systems Laboratory Testing Nearly Complete

    NASA Technical Reports Server (NTRS)

    2005-01-01

    Radio history continues to be made at the NASA Lewis Research Center with the completion of phase one of the digital audio radio (DAR) testing conducted by the Consumer Electronics Group of the Electronic Industries Association. This satellite, satellite/terrestrial, and terrestrial digital technology will open up new audio broadcasting opportunities both domestically and worldwide. It will significantly improve the current quality of amplitude-modulated/frequency-modulated (AM/FM) radio with a new digitally modulated radio signal and will introduce true compact-disc-quality (CD-quality) sound for the first time. Lewis is hosting the laboratory testing of seven proposed digital audio radio systems and modes. Two of the proposed systems operate in two modes each, making a total of nine systems being tested. The nine systems are divided into the following types of transmission: in-band on-channel (IBOC), in-band adjacent-channel (IBAC), and new bands. The laboratory testing was conducted by the Consumer Electronics Group of the Electronic Industries Association. Subjective assessments of the audio recordings for each of the nine systems was conducted by the Communications Research Center in Ottawa, Canada, under contract to the Electronic Industries Association. The Communications Research Center has the only CCIR-qualified (Consultative Committee for International Radio) audio testing facility in North America. The main goals of the U.S. testing process are to (1) provide technical data to the Federal Communication Commission (FCC) so that it can establish a standard for digital audio receivers and transmitters and (2) provide the receiver and transmitter industries with the proper standards upon which to build their equipment. In addition, the data will be forwarded to the International Telecommunications Union to help in the establishment of international standards for digital audio receivers and transmitters, thus allowing U.S. manufacturers to compete in the

  13. Retinotopic effects during spatial audio-visual integration.

    PubMed

    Meienbrock, A; Naumer, M J; Doehrmann, O; Singer, W; Muckli, L

    2007-02-01

    The successful integration of visual and auditory stimuli requires information about whether visual and auditory signals originate from corresponding places in the external world. Here we report crossmodal effects of spatially congruent and incongruent audio-visual (AV) stimulation. Visual and auditory stimuli were presented from one of four horizontal locations in external space. Seven healthy human subjects had to assess the spatial fit of a visual stimulus (i.e. a gray-scaled picture of a cartoon dog) and a simultaneously presented auditory stimulus (i.e. a barking sound). Functional magnetic resonance imaging (fMRI) revealed two distinct networks of cortical regions that processed preferentially either spatially congruent or spatially incongruent AV stimuli. Whereas earlier visual areas responded preferentially to incongruent AV stimulation, higher visual areas of the temporal and parietal cortex (left inferior temporal gyrus [ITG], right posterior superior temporal gyrus/sulcus [pSTG/STS], left intra-parietal sulcus [IPS]) and frontal regions (left pre-central gyrus [PreCG], left dorsolateral pre-frontal cortex [DLPFC]) responded preferentially to congruent AV stimulation. A position-resolved analysis revealed three robust cortical representations for each of the four visual stimulus locations in retinotopic visual regions corresponding to the representation of the horizontal meridian in area V1 and at the dorsal and ventral borders between areas V2 and V3. While these regions of interest (ROIs) did not show any significant effect of spatial congruency, we found subregions within ROIs in the right hemisphere that showed an incongruency effect (i.e. an increased fMRI signal during spatially incongruent compared to congruent AV stimulation). We interpret this finding as a correlate of spatially distributed recurrent feedback during mismatch processing: whenever a spatial mismatch is detected in multisensory regions (such as the IPS), processing resources are re

  14. GK Per (Nova Persei 1901): HUBBLE SPACE TELESCOPE IMAGERY AND SPECTROSCOPY OF THE EJECTA, AND FIRST SPECTRUM OF THE JET-LIKE FEATURE

    SciTech Connect

    Shara, Michael M.; Zurek, David; Mizusawa, Trisha; De Marco, Orsola; Williams, Robert; Livio, Mario

    2012-06-15

    We have imaged the ejecta of GK Persei (Nova Persei 1901 A.D.) with the Hubble Space Telescope (HST), whose 0.1 arcsec resolution reveals hundreds of cometary-like structures with long axes aligned toward GK Per. One or both ends of the structures often show a brightness enhancement relative to the structures' middle sections, but there is no simple regularity to their morphologies (in contrast with, for example, the Helix nebula). Some of structures' morphologies suggest the presence of slow-moving or stationary material with which the ejecta is colliding, while others suggest shaping from a wind emanating from GK Per itself. The most detailed expansion map of any classical nova's ejecta was created by comparing HST images taken in successive years. Wide Field and Planetary Camera 2 narrowband images and Space Telescope Imaging Spectrograph spectra demonstrate that the physical conditions in this nova's ejecta vary strongly on spatial scales much smaller than those of the ejecta. Directly measuring accurate densities and compositions, and hence masses of this and other nova shells, will demand data at least as resolved spatially as those presented here. The filling factor of the ejecta is 1% or less, and the nova ejecta mass must be less than 10{sup -4} M{sub Sun }. A modest fraction of the emission nebulosities vary in brightness by up to a factor of two on timescales of one year. Finally, we present the deepest images yet obtained of a jet-like feature outside the main body of GK Per nebulosity, and the first spectrum of that feature. Dominated by strong, narrow emission lines of [N II], [O II], [O III], and [S II], this feature is probably a shock due to ejected material running into stationary interstellar matter, slowly moving ejecta from a previous nova episode, or circumbinary matter present before 1901. An upper limit to the mass of the jet is of order a few times 10{sup -6} M{sub Sun }. If the jet mass is close to this limit then the jet might be an

  15. Talker variability in audio-visual speech perception

    PubMed Central

    Heald, Shannon L. M.; Nusbaum, Howard C.

    2014-01-01

    A change in talker is a change in the context for the phonetic interpretation of acoustic patterns of speech. Different talkers have different mappings between acoustic patterns and phonetic categories and listeners need to adapt to these differences. Despite this complexity, listeners are adept at comprehending speech in multiple-talker contexts, albeit at a slight but measurable performance cost (e.g., slower recognition). So far, this talker variability cost has been demonstrated only in audio-only speech. Other research in single-talker contexts have shown, however, that when listeners are able to see a talker’s face, speech recognition is improved under adverse listening (e.g., noise or distortion) conditions that can increase uncertainty in the mapping between acoustic patterns and phonetic categories. Does seeing a talker’s face reduce the cost of word recognition in multiple-talker contexts? We used a speeded word-monitoring task in which listeners make quick judgments about target word recognition in single- and multiple-talker contexts. Results show faster recognition performance in single-talker conditions compared to multiple-talker conditions for both audio-only and audio-visual speech. However, recognition time in a multiple-talker context was slower in the audio-visual condition compared to audio-only condition. These results suggest that seeing a talker’s face during speech perception may slow recognition by increasing the importance of talker identification, signaling to the listener a change in talker has occurred. PMID:25076919

  16. Multifunction audio digitizer for communications systems

    NASA Technical Reports Server (NTRS)

    Monford, L. G., Jr.

    1971-01-01

    Digitizer accomplishes both N bit pulse code modulation /PCM/ and delta modulation, and provides modulation indicating variable signal gain and variable sidetone. Other features include - low package count, variable clock rate to optimize bandwidth, and easily expanded PCM output.

  17. Why 8-Year-Olds Cannot Tell the Difference between Steve Martin and Paul Newman: Factors Contributing to the Slow Development of Sensitivity to the Spacing of Facial Features

    ERIC Educational Resources Information Center

    Mondloch, Catherine J.; Dobson, Kate S.; Parsons, Julie; Maurer, Daphne

    2004-01-01

    Children are nearly as sensitive as adults to some cues to facial identity (e.g., differences in the shape of internal features and the external contour), but children are much less sensitive to small differences in the spacing of facial features. To identify factors that contribute to this pattern, we compared 8-year-olds' sensitivity to spacing…

  18. FIRST ULTRAVIOLET REFLECTANCE SPECTRA OF PLUTO AND CHARON BY THE HUBBLE SPACE TELESCOPE COSMIC ORIGINS SPECTROGRAPH: DETECTION OF ABSORPTION FEATURES AND EVIDENCE FOR TEMPORAL CHANGE

    SciTech Connect

    Stern, S. A.; Spencer, J. R.; Shinn, A.; Cunningham, N. J.; Hain, M. J.

    2012-01-15

    We have observed the mid-UV spectra of both Pluto and its large satellite, Charon, at two rotational epochs using the Hubble Space Telescope (HST) Cosmic Origins Spectrograph (COS) in 2010. These are the first HST/COS measurements of Pluto and Charon. Here we describe the observations and our reduction of them, and present the albedo spectra, average mid-UV albedos, and albedo slopes we derive from these data. These data reveal evidence for a strong absorption feature in the mid-UV spectrum of Pluto; evidence for temporal change in Pluto's spectrum since the 1990s is reported, and indirect evidence for a near-UV spectral absorption on Charon is also reported.

  19. #NASATweetup @NASA_Langley Audio Slideshow

    NASA Video Gallery

    NASA Langley Research Center's first tweet-up involved a diverse group of more than 40 that included an astronaut's daughter, a physics student from Wisconsin, one of NASA's newest space camp crew ...

  20. Say What? The Role of Audio in Multimedia Video

    NASA Astrophysics Data System (ADS)

    Linder, C. A.; Holmes, R. M.

    2011-12-01

    Audio, including interviews, ambient sounds, and music, is a critical-yet often overlooked-part of an effective multimedia video. In February 2010, Linder joined scientists working on the Global Rivers Observatory Project for two weeks of intensive fieldwork in the Congo River watershed. The team's goal was to learn more about how climate change and deforestation are impacting the river system and coastal ocean. Using stills and video shot with a lightweight digital SLR outfit and audio recorded with a pocket-sized sound recorder, Linder documented the trials and triumphs of working in the heart of Africa. Using excerpts from the six-minute Congo multimedia video, this presentation will illustrate how to record and edit an engaging audio track. Topics include interview technique, collecting ambient sounds, choosing and using music, and editing it all together to educate and entertain the viewer.

  1. Highlight summarization in golf videos using audio signals

    NASA Astrophysics Data System (ADS)

    Kim, Hyoung-Gook; Kim, Jin Young

    2008-01-01

    In this paper, we present an automatic summarization of highlights in golf videos based on audio information alone without video information. The proposed highlight summarization system is carried out based on semantic audio segmentation and detection on action units from audio signals. Studio speech, field speech, music, and applause are segmented by means of sound classification. Swing is detected by the methods of impulse onset detection. Sounds like swing and applause form a complete action unit, while studio speech and music parts are used to anchor the program structure. With the advantage of highly precise detection of applause, highlights are extracted effectively. Our experimental results obtain high classification precision on 18 golf games. It proves that the proposed system is very effective and computationally efficient to apply the technology to embedded consumer electronic devices.

  2. Music Identification System Using MPEG-7 Audio Signature Descriptors

    PubMed Central

    You, Shingchern D.; Chen, Wei-Hwa; Chen, Woei-Kae

    2013-01-01

    This paper describes a multiresolution system based on MPEG-7 audio signature descriptors for music identification. Such an identification system may be used to detect illegally copied music circulated over the Internet. In the proposed system, low-resolution descriptors are used to search likely candidates, and then full-resolution descriptors are used to identify the unknown (query) audio. With this arrangement, the proposed system achieves both high speed and high accuracy. To deal with the problem that a piece of query audio may not be inside the system's database, we suggest two different methods to find the decision threshold. Simulation results show that the proposed method II can achieve an accuracy of 99.4% for query inputs both inside and outside the database. Overall, it is highly possible to use the proposed system for copyright control. PMID:23533359

  3. Multi-channel spatialization system for audio signals

    NASA Technical Reports Server (NTRS)

    Begault, Durand R. (Inventor)

    1995-01-01

    Synthetic head related transfer functions (HRTF's) for imposing reprogramable spatial cues to a plurality of audio input signals included, for example, in multiple narrow-band audio communications signals received simultaneously are generated and stored in interchangeable programmable read only memories (PROM's) which store both head related transfer function impulse response data and source positional information for a plurality of desired virtual source locations. The analog inputs of the audio signals are filtered and converted to digital signals from which synthetic head related transfer functions are generated in the form of linear phase finite impulse response filters. The outputs of the impulse response filters are subsequently reconverted to analog signals, filtered, mixed and fed to a pair of headphones.

  4. Multi-channel spatialization systems for audio signals

    NASA Technical Reports Server (NTRS)

    Begault, Durand R. (Inventor)

    1993-01-01

    Synthetic head related transfer functions (HRTF's) for imposing reprogrammable spatial cues to a plurality of audio input signals included, for example, in multiple narrow-band audio communications signals received simultaneously are generated and stored in interchangeable programmable read only memories (PROM's) which store both head related transfer function impulse response data and source positional information for a plurality of desired virtual source locations. The analog inputs of the audio signals are filtered and converted to digital signals from which synthetic head related transfer functions are generated in the form of linear phase finite impulse response filters. The outputs of the impulse response filters are subsequently reconverted to analog signals, filtered, mixed, and fed to a pair of headphones.

  5. The Application of Acoustic Measurements and Audio Recordings for Diagnosis of In-Flight Hardware Anomalies

    NASA Technical Reports Server (NTRS)

    Welsh, David; Denham, Samuel; Allen, Christopher

    2011-01-01

    In many cases, an initial symptom of hardware malfunction is unusual or unexpected acoustic noise. Many industries such as automotive, heating and air conditioning, and petro-chemical processing use noise and vibration data along with rotating machinery analysis techniques to identify noise sources and correct hardware defects. The NASA/Johnson Space Center Acoustics Office monitors the acoustic environment of the International Space Station (ISS) through periodic sound level measurement surveys. Trending of the sound level measurement survey results can identify in-flight hardware anomalies. The crew of the ISS also serves as a "detection tool" in identifying unusual hardware noises; in these cases the spectral analysis of audio recordings made on orbit can be used to identify hardware defects that are related to rotating components such as fans, pumps, and compressors. In this paper, three examples of the use of sound level measurements and audio recordings for the diagnosis of in-flight hardware anomalies are discussed: identification of blocked inter-module ventilation (IMV) ducts, diagnosis of abnormal ISS Crew Quarters rack exhaust fan noise, and the identification and replacement of a defective flywheel assembly in the Treadmill with Vibration Isolation (TVIS) hardware. In each of these examples, crew time was saved by identifying the off nominal component or condition that existed and in directing in-flight maintenance activities to address and correct each of these problems.

  6. Video-assisted segmentation of speech and audio track

    NASA Astrophysics Data System (ADS)

    Pandit, Medha; Yusoff, Yusseri; Kittler, Josef; Christmas, William J.; Chilton, E. H. S.

    1999-08-01

    Video database research is commonly concerned with the storage and retrieval of visual information invovling sequence segmentation, shot representation and video clip retrieval. In multimedia applications, video sequences are usually accompanied by a sound track. The sound track contains potential cues to aid shot segmentation such as different speakers, background music, singing and distinctive sounds. These different acoustic categories can be modeled to allow for an effective database retrieval. In this paper, we address the problem of automatic segmentation of audio track of multimedia material. This audio based segmentation can be combined with video scene shot detection in order to achieve partitioning of the multimedia material into semantically significant segments.

  7. Evaluation of robustness and transparency of multiple audio watermark embedding

    NASA Astrophysics Data System (ADS)

    Steinebach, Martin; Zmudzinski, Sascha

    2008-02-01

    As digital watermarking becomes an accepted and widely applied technology, a number of concerns regarding its reliability in typical application scenarios come up. One important and often discussed question is the robustness of digital watermarks against multiple embedding. This means that one cover is marked several times by various users with by same watermarking algorithm but with different keys and different watermark messages. In our paper we discuss the behavior of our PCM audio watermarking algorithm when applying multiple watermark embedding. This includes evaluation of robustness and transparency. Test results for multiple hours of audio content ranging from spoken words to music are provided.

  8. Animation, audio, and spatial ability: Optimizing multimedia for scientific explanations

    NASA Astrophysics Data System (ADS)

    Koroghlanian, Carol May

    This study investigated the effects of audio, animation and spatial ability in a computer based instructional program for biology. The program presented instructional material via text or audio with lean text and included eight instructional sequences presented either via static illustrations or animations. High school students enrolled in a biology course were blocked by spatial ability and randomly assigned to one of four treatments (Text-Static Illustration Audio-Static Illustration, Text-Animation, Audio-Animation). The study examined the effects of instructional mode (Text vs. Audio), illustration mode (Static Illustration vs. Animation) and spatial ability (Low vs. High) on practice and posttest achievement, attitude and time. Results for practice achievement indicated that high spatial ability participants achieved more than low spatial ability participants. Similar results for posttest achievement and spatial ability were not found. Participants in the Static Illustration treatments achieved the same as participants in the Animation treatments on both the practice and posttest. Likewise, participants in the Text treatments achieved the same as participants in the Audio treatments on both the practice and posttest. In terms of attitude, participants responded favorably to the computer based instructional program. They found the program interesting, felt the static illustrations or animations made the explanations easier to understand and concentrated on learning the material. Furthermore, participants in the Animation treatments felt the information was easier to understand than participants in the Static Illustration treatments. However, no difference for any attitude item was found for participants in the Text as compared to those in the Audio treatments. Significant differences were found by Spatial Ability for three attitude items concerning concentration and interest. In all three items, the low spatial ability participants responded more positively

  9. Three dimensional audio versus head down TCAS displays

    NASA Technical Reports Server (NTRS)

    Begault, Durand R.; Pittman, Marc T.

    1994-01-01

    The advantage of a head up auditory display was evaluated in an experiment designed to measure and compare the acquisition time for capturing visual targets under two conditions: Standard head down traffic collision avoidance system (TCAS) display, and three-dimensional (3-D) audio TCAS presentation. Ten commercial airline crews were tested under full mission simulation conditions at the NASA Ames Crew-Vehicle Systems Research Facility Advanced Concepts Flight Simulator. Scenario software generated targets corresponding to aircraft which activated a 3-D aural advisory or a TCAS advisory. Results showed a significant difference in target acquisition time between the two conditions, favoring the 3-D audio TCAS condition by 500 ms.

  10. Cover signal specific steganalysis: the impact of training on the example of two selected audio steganalysis approaches

    NASA Astrophysics Data System (ADS)

    Kraetzer, Christian; Dittmann, Jana

    2008-02-01

    The main goals of this paper are to show the impact of the basic assumptions for the cover channel characteristics as well as the impact of different training/testing set generation strategies on the statistical detectability of exemplary chosen audio hiding approaches known from steganography and watermarking. Here we have selected exemplary five steganography algorithms and four watermarking algorithms. The channel characteristics for two different chosen audio cover channels (an application specific exemplary scenario of VoIP steganography and universal audio steganography) are formalised and their impact on decisions in the steganalysis process, especially on the strategies applied for training/ testing set generation, are shown. Following the assumptions on the cover channel characteristics either cover dependent or cover independent training and testing can be performed, using either correlated or non-correlated training and test sets. In comparison to previous work, additional frequency domain features are introduced for steganalysis and the performance (in terms of classification accuracy) of Bayesian classifiers and multinomial logistic regression models is compared with the results of SVM classification. We show that the newly implemented frequency domain features increase the classification accuracy achieved in SVM classification. Furthermore it is shown on the example of VoIP steganalysis that channel character specific evaluation performs better than tests without focus on a specific channel (i.e. universal steganalysis). A comparison of test results for cover dependent and independent training and testing shows that the latter performs better for all nine algorithms evaluated here and the used SVM based classifier.

  11. Multi-modal gesture recognition using integrated model of motion, audio and video

    NASA Astrophysics Data System (ADS)

    Goutsu, Yusuke; Kobayashi, Takaki; Obara, Junya; Kusajima, Ikuo; Takeichi, Kazunari; Takano, Wataru; Nakamura, Yoshihiko

    2015-07-01

    Gesture recognition is used in many practical applications such as human-robot interaction, medical rehabilitation and sign language. With increasing motion sensor development, multiple data sources have become available, which leads to the rise of multi-modal gesture recognition. Since our previous approach to gesture recognition depends on a unimodal system, it is difficult to classify similar motion patterns. In order to solve this problem, a novel approach which integrates motion, audio and video models is proposed by using dataset captured by Kinect. The proposed system can recognize observed gestures by using three models. Recognition results of three models are integrated by using the proposed framework and the output becomes the final result. The motion and audio models are learned by using Hidden Markov Model. Random Forest which is the video classifier is used to learn the video model. In the experiments to test the performances of the proposed system, the motion and audio models most suitable for gesture recognition are chosen by varying feature vectors and learning methods. Additionally, the unimodal and multi-modal models are compared with respect to recognition accuracy. All the experiments are conducted on dataset provided by the competition organizer of MMGRC, which is a workshop for Multi-Modal Gesture Recognition Challenge. The comparison results show that the multi-modal model composed of three models scores the highest recognition rate. This improvement of recognition accuracy means that the complementary relationship among three models improves the accuracy of gesture recognition. The proposed system provides the application technology to understand human actions of daily life more precisely.

  12. Fiber-channel audio video standard for military and commercial aircraft product lines

    NASA Astrophysics Data System (ADS)

    Keller, Jack E.

    2002-08-01

    Fibre channel is an emerging high-speed digital network technology that combines to make inroads into the avionics arena. The suitability of fibre channel for such applications is largely due to its flexibility in these key areas: Network topologies can be configured in point-to-point, arbitrated loop or switched fabric connections. The physical layer supports either copper or fiber optic implementations with a Bit Error Rate of less than 10-12. Multiple Classes of Service are available. Multiple Upper Level Protocols are supported. Multiple high speed data rates offer open ended growth paths providing speed negotiation within a single network. Current speeds supported by commercially available hardware are 1 and 2 Gbps providing effective data rates of 100 and 200 MBps respectively. Such networks lend themselves well to the transport of digital video and audio data. This paper summarizes an ANSI standard currently in the final approval cycle of the InterNational Committee for Information Technology Standardization (INCITS). This standard defines a flexible mechanism whereby digital video, audio and ancillary data are systematically packaged for transport over a fibre channel network. The basic mechanism, called a container, houses audio and video content functionally grouped as elements of the container called objects. Featured in this paper is a specific container mapping called Simple Parametric Digital Video (SPDV) developed particularly to address digital video in avionics systems. SPDV provides pixel-based video with associated ancillary data typically sourced by various sensors to be processed and/or distributed in the cockpit for presentation via high-resolution displays. Also highlighted in this paper is a streamlined Upper Level Protocol (ULP) called Frame Header Control Procedure (FHCP) targeted for avionics systems where the functionality of a more complex ULP is not required.

  13. Choosing efficient feature sets for video classification

    NASA Astrophysics Data System (ADS)

    Fischer, Stephan; Steinmetz, Ralf

    1998-12-01

    In this paper, we address the problem of choosing appropriate features to describe the content of still pictures or video sequences, including audio. As the computational analysis of these features is often time- consuming, it is useful to identify a minimal set allowing for an automatic classification of some class or genre. Further, it can be shown that deleting the coherence of the features characterizing some class, is not suitable to guarantee an optimal classification result. The central question of the paper is thus, which features should be selected, and how they should be weighted to optimize a classification problem.

  14. Survey of the State of Audio Collections in Academic Libraries

    ERIC Educational Resources Information Center

    Smith, Abby; Allen, David Randal; Allen, Karen

    2004-01-01

    The goal of this survey was to collect and analyze baseline information about the status of audio collections held by a set of research institutions. This information can help shape the national preservation plan now being developed by the National Recording Preservation Board (NRPB) and the Library of Congress to preserve "sound recordings that…

  15. Audio and Video Reflections to Promote Social Justice

    ERIC Educational Resources Information Center

    Boske, Christa

    2011-01-01

    Purpose: The purpose of this paper is to examine how 15 graduate students enrolled in a US school leadership preparation program understand issues of social justice and equity through a reflective process utilizing audio and/or video software. Design/methodology/approach: The study is based on the tradition of grounded theory. The researcher…

  16. The Audio-Visual Marketing Handbook for Independent Schools.

    ERIC Educational Resources Information Center

    Griffith, Tom

    This how-to booklet offers specific advice on producing video or slide/tape programs for marketing independent schools. Five chapters present guidelines for various stages in the process: (1) Audio-Visual Marketing in Context (aesthetics and economics of audiovisual marketing); (2) A Question of Identity (identifying the audience and deciding on…

  17. An Evaluation of the Audio Workbook System. R & D Report.

    ERIC Educational Resources Information Center

    Andrulis, Richard S.

    The Cassette Review Program (CRP), developed by The American College of Life Underwriters, is organized into 10 sections corresponding to the 10 courses of the American College C.L.U. diploma program. It includes both audio tapes and notebooks. The formative evaluation of the CRP carried out in 1971 resulted in a restructuring of both the tapes…

  18. Audio Visual Technology and the Teaching of Foreign Languages.

    ERIC Educational Resources Information Center

    Halbig, Michael C.

    Skills in comprehending the spoken language source are becoming increasingly important due to the audio-visual orientation of our culture. It would seem natural, therefore, to adjust the learning goals and environment accordingly. The video-cassette machine is an ideal means for creating this learning environment and developing the listening…

  19. The 7 Habits of Highly Effective Families Audio System. [Audiotapes].

    ERIC Educational Resources Information Center

    Covey, Stephen R.

    Intended to help families build rewarding relationships, this set of audio tapes communicates the importance of family, shows the role of leadership in creating "beautiful family culture," and provides a frame of reference in which to solve problems. The first of the four tapes explores the concept of principles as the foundation of family, and…

  20. Integrated Spacesuit Audio System Enhances Speech Quality and Reduces Noise

    NASA Technical Reports Server (NTRS)

    Huang, Yiteng Arden; Chen, Jingdong; Chen, Shaoyan Sharyl

    2009-01-01

    A new approach has been proposed for increasing astronaut comfort and speech capture. Currently, the special design of a spacesuit forms an extreme acoustic environment making it difficult to capture clear speech without compromising comfort. The proposed Integrated Spacesuit Audio (ISA) system is to incorporate the microphones into the helmet and use software to extract voice signals from background noise.

  1. Exploratory Evaluation of Audio Email Technology in Formative Assessment Feedback

    ERIC Educational Resources Information Center

    Macgregor, George; Spiers, Alex; Taylor, Chris

    2011-01-01

    Formative assessment generates feedback on students' performance, thereby accelerating and improving student learning. Anecdotal evidence gathered by a number of evaluations has hypothesised that audio feedback may be capable of enhancing student learning more than other approaches. In this paper we report on the preliminary findings of a…

  2. Ultrahigh and audio frequencies in a laser beam

    SciTech Connect

    Casabella, P.A.; Gonsiorowski, T.; Leitner, A.

    1980-05-01

    The helium--neon lasers readily available in teaching laboratories usually operate in several photon modes simultaneously. The first-difference and second-difference beats lie in the uhf- and audio-frequency ranges, respectively, and can be detected as sinusoidal signals with photodiodes. These are instructive experiments which raise thought provoking questions about cavity resonance and negative dispersion.

  3. A Decade of Experimentation with the Audio-Tutorial System.

    ERIC Educational Resources Information Center

    Mott, Charles J.

    The audio-tutorial systems in earth science was developed at St. Petersburg Junior College in response to an unacceptable but not atypical student failure rate at the community college level. This description of the program includes: (1) history of program development, (2) use of student tutors, (3) experimentation with a student paced, mastery…

  4. National Center for Audio Tapes 1970-72 Catalog.

    ERIC Educational Resources Information Center

    1970

    Over 12,000 audio tapes are listed in this catalog under one of seven headings: arts, education, language and literature, mathematics-science, physical-recreational activities, social studies, and vocational-technical training. Information on each tape includes its Library of Congress subject classification, grade level, series title and…

  5. National Center for Audio Tapes 1971 Catalog Supplement.

    ERIC Educational Resources Information Center

    Colorado Univ., Boulder. National Center for Audio Tapes.

    About 600 audio tapes are cataloged in this supplement to the 1970-71 catalog (ED 038 854). The catalog is organized in three sections. The subject index, based on the Library of Congress subject area classification scheme, includes these categories: art, education, languages and literature, mathematics, physical education and recreation, science,…

  6. Influence of Audio-Visual Presentations on Learning Abstract Concepts.

    ERIC Educational Resources Information Center

    Lai, Shu-Ling

    2000-01-01

    Describes a study of college students that investigated whether various types of visual illustrations influenced abstract concept learning when combined with audio instruction. Discusses results of analysis of variance and pretest posttest scores in relation to learning performance, attitudes toward the computer-based program, and differences in…

  7. Sounds Good: Using Digital Audio for Evaluation Feedback

    ERIC Educational Resources Information Center

    Rotheram, Bob

    2009-01-01

    Feedback on student work is problematic for faculty and students in British higher education. Evaluation feedback takes faculty much time to produce and students are often dissatisfied with its quantity, timing, and clarity. The Sounds Good project has been experimenting with the use of digital audio for feedback, aiming to save faculty time and…

  8. Audio-Described Educational Materials: Ugandan Teachers' Experiences

    ERIC Educational Resources Information Center

    Wormnaes, Siri; Sellaeg, Nina

    2013-01-01

    This article describes and discusses a qualitative, descriptive, and exploratory study of how 12 visually impaired teachers in Uganda experienced audio-described educational video material for teachers and student teachers. The study is based upon interviews with these teachers and observations while they were using the material either…

  9. Australian High Schools Use Audio-Tutorials in Field Biology

    ERIC Educational Resources Information Center

    Meyer, G. R.; Postlethwait, S. N.

    1970-01-01

    Describes a workshop program for biology teachers involving the preparation of audio-tutorial units designed to prepare high school students for independent study of field ecology. Outlines units prepared and discusses usefulness of similar units in high schools and elsewhere. (EB)

  10. Digital Audio Broadcasting in the Short Wave Bands

    NASA Technical Reports Server (NTRS)

    Vaisnys, Arvydas

    1998-01-01

    For many decades the Short Wae broadcasting service has used high power, double-sideband AM signals to reach audiences far and wide. While audio quality was usually not very high, inexpensive receivers could be used to tune into broadcasts fro distant countries.

  11. Developing a Framework for Effective Audio Feedback: A Case Study

    ERIC Educational Resources Information Center

    Hennessy, Claire; Forrester, Gillian

    2014-01-01

    The increase in the use of technology-enhanced learning in higher education has included a growing interest in new approaches to enhance the quality of feedback given to students. Audio feedback is one method that has become more popular, yet evaluating its role in feedback delivery is still an emerging area for research. This paper is based on a…

  12. Audio-Visual Aid in Teaching "Fatty Liver"

    ERIC Educational Resources Information Center

    Dash, Sambit; Kamath, Ullas; Rao, Guruprasad; Prakash, Jay; Mishra, Snigdha

    2016-01-01

    Use of audio visual tools to aid in medical education is ever on a rise. Our study intends to find the efficacy of a video prepared on "fatty liver," a topic that is often a challenge for pre-clinical teachers, in enhancing cognitive processing and ultimately learning. We prepared a video presentation of 11:36 min, incorporating various…

  13. Audio-visual perception system for a humanoid robotic head.

    PubMed

    Viciana-Abad, Raquel; Marfil, Rebeca; Perez-Lorenzo, Jose M; Bandera, Juan P; Romero-Garces, Adrian; Reche-Lopez, Pedro

    2014-01-01

    One of the main issues within the field of social robotics is to endow robots with the ability to direct attention to people with whom they are interacting. Different approaches follow bio-inspired mechanisms, merging audio and visual cues to localize a person using multiple sensors. However, most of these fusion mechanisms have been used in fixed systems, such as those used in video-conference rooms, and thus, they may incur difficulties when constrained to the sensors with which a robot can be equipped. Besides, within the scope of interactive autonomous robots, there is a lack in terms of evaluating the benefits of audio-visual attention mechanisms, compared to only audio or visual approaches, in real scenarios. Most of the tests conducted have been within controlled environments, at short distances and/or with off-line performance measurements. With the goal of demonstrating the benefit of fusing sensory information with a Bayes inference for interactive robotics, this paper presents a system for localizing a person by processing visual and audio data. Moreover, the performance of this system is evaluated and compared via considering the technical limitations of unimodal systems. The experiments show the promise of the proposed approach for the proactive detection and tracking of speakers in a human-robot interactive framework. PMID:24878593

  14. Improved Techniques for Automatic Chord Recognition from Music Audio Signals

    ERIC Educational Resources Information Center

    Cho, Taemin

    2014-01-01

    This thesis is concerned with the development of techniques that facilitate the effective implementation of capable automatic chord transcription from music audio signals. Since chord transcriptions can capture many important aspects of music, they are useful for a wide variety of music applications and also useful for people who learn and perform…

  15. Audio Card Systems. Technical Information Bulletin No. 13.

    ERIC Educational Resources Information Center

    Gasser, P.

    This examination of audio card systems for computers begins by identifying the three information processing systems for sound: sound digitizing, synthesis of text, and word recognition. Specific pedagogical applications of digitized sound are then briefly discussed. The remainder of the document focuses on specifications for the working of vocal…

  16. The Timbre Toolbox: extracting audio descriptors from musical signals.

    PubMed

    Peeters, Geoffroy; Giordano, Bruno L; Susini, Patrick; Misdariis, Nicolas; McAdams, Stephen

    2011-11-01

    The analysis of musical signals to extract audio descriptors that can potentially characterize their timbre has been disparate and often too focused on a particular small set of sounds. The Timbre Toolbox provides a comprehensive set of descriptors that can be useful in perceptual research, as well as in music information retrieval and machine-learning approaches to content-based retrieval in large sound databases. Sound events are first analyzed in terms of various input representations (short-term Fourier transform, harmonic sinusoidal components, an auditory model based on the equivalent rectangular bandwidth concept, the energy envelope). A large number of audio descriptors are then derived from each of these representations to capture temporal, spectral, spectrotemporal, and energetic properties of the sound events. Some descriptors are global, providing a single value for the whole sound event, whereas others are time-varying. Robust descriptive statistics are used to characterize the time-varying descriptors. To examine the information redundancy across audio descriptors, correlational analysis followed by hierarchical clustering is performed. This analysis suggests ten classes of relatively independent audio descriptors, showing that the Timbre Toolbox is a multidimensional instrument for the measurement of the acoustical structure of complex sound signals. PMID:22087919

  17. Effectiveness of Audio on Screen Captures in Software Application Instruction

    ERIC Educational Resources Information Center

    Veronikas, Susan Walsh; Maushak, Nancy

    2005-01-01

    Presentation of software instruction has been supported by manuals and textbooks consisting of screen captures, but a multimedia approach may increase learning outcomes. This study investigated the effects of modality (text, audio, or dual) on the achievement and attitudes of college students learning a software application through the computer.…

  18. Infant Perception of Audio-Visual Speech Synchrony

    ERIC Educational Resources Information Center

    Lewkowicz, David J.

    2010-01-01

    Three experiments investigated perception of audio-visual (A-V) speech synchrony in 4- to 10-month-old infants. Experiments 1 and 2 used a convergent-operations approach by habituating infants to an audiovisually synchronous syllable (Experiment 1) and then testing for detection of increasing degrees of A-V asynchrony (366, 500, and 666 ms) or by…

  19. Audio-Visual Perception System for a Humanoid Robotic Head

    PubMed Central

    Viciana-Abad, Raquel; Marfil, Rebeca; Perez-Lorenzo, Jose M.; Bandera, Juan P.; Romero-Garces, Adrian; Reche-Lopez, Pedro

    2014-01-01

    One of the main issues within the field of social robotics is to endow robots with the ability to direct attention to people with whom they are interacting. Different approaches follow bio-inspired mechanisms, merging audio and visual cues to localize a person using multiple sensors. However, most of these fusion mechanisms have been used in fixed systems, such as those used in video-conference rooms, and thus, they may incur difficulties when constrained to the sensors with which a robot can be equipped. Besides, within the scope of interactive autonomous robots, there is a lack in terms of evaluating the benefits of audio-visual attention mechanisms, compared to only audio or visual approaches, in real scenarios. Most of the tests conducted have been within controlled environments, at short distances and/or with off-line performance measurements. With the goal of demonstrating the benefit of fusing sensory information with a Bayes inference for interactive robotics, this paper presents a system for localizing a person by processing visual and audio data. Moreover, the performance of this system is evaluated and compared via considering the technical limitations of unimodal systems. The experiments show the promise of the proposed approach for the proactive detection and tracking of speakers in a human-robot interactive framework. PMID:24878593

  20. Audio-Visual Communications, A Tool for the Professional

    ERIC Educational Resources Information Center

    Journal of Environmental Health, 1976

    1976-01-01

    The manner in which the Cuyahoga County, Ohio Department of Environmental Health utilizes audio-visual presentations for communication with business and industry, professional public health agencies and the general public is presented. Subjects including food sanitation, radiation protection and safety are described. (BT)

  1. Parametric Packet-Layer Model for Evaluation Audio Quality in Multimedia Streaming Services

    NASA Astrophysics Data System (ADS)

    Egi, Noritsugu; Hayashi, Takanori; Takahashi, Akira

    We propose a parametric packet-layer model for monitoring audio quality in multimedia streaming services such as Internet protocol television (IPTV). This model estimates audio quality of experience (QoE) on the basis of quality degradation due to coding and packet loss of an audio sequence. The input parameters of this model are audio bit rate, sampling rate, frame length, packet-loss frequency, and average burst length. Audio bit rate, packet-loss frequency, and average burst length are calculated from header information in received IP packets. For sampling rate, frame length, and audio codec type, the values or the names used in monitored services are input into this model directly. We performed a subjective listening test to examine the relationships between these input parameters and perceived audio quality. The codec used in this test was the Advanced Audio Codec-Low Complexity (AAC-LC), which is one of the international standards for audio coding. On the basis of the test results, we developed an audio quality evaluation model. The verification results indicate that audio quality estimated by the proposed model has a high correlation with perceived audio quality.

  2. Hearing You Loud and Clear: Student Perspectives of Audio Feedback in Higher Education

    ERIC Educational Resources Information Center

    Gould, Jill; Day, Pat

    2013-01-01

    The use of audio feedback for students in a full-time community nursing degree course is appraised. The aim of this mixed methods study was to examine student views on audio feedback for written assignments. Questionnaires and a focus group were used to capture student opinion of this pilot project. The majority of students valued audio feedback…

  3. Audio Use in E-Learning: What, Why, When, and How?

    ERIC Educational Resources Information Center

    Calandra, Brendan; Barron, Ann E.; Thompson-Sellers, Ingrid

    2008-01-01

    Decisions related to the implementation of audio in e-learning are perplexing for many instructional designers, and deciphering theory and principles related to audio use can be difficult for practitioners. Yet, as bandwidth on the Internet increases, digital audio is becoming more common in online courses. This article provides a review of…

  4. Responding Effectively to Composition Students: Comparing Student Perceptions of Written and Audio Feedback

    ERIC Educational Resources Information Center

    Bilbro, J.; Iluzada, C.; Clark, D. E.

    2013-01-01

    The authors compared student perceptions of audio and written feedback in order to assess what types of students may benefit from receiving audio feedback on their essays rather than written feedback. Many instructors previously have reported the advantages they see in audio feedback, but little quantitative research has been done on how the…

  5. 78 FR 18416 - Sixth Meeting: RTCA Special Committee 226, Audio Systems and Equipment

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-03-26

    ... Federal Aviation Administration Sixth Meeting: RTCA Special Committee 226, Audio Systems and Equipment... Notice of RTCA Special Committee 226, Audio Systems and Equipment. SUMMARY: The FAA is issuing this notice to advise the public of the sixth meeting of the RTCA Special Committee 226, Audio Systems...

  6. 78 FR 38093 - Seventh Meeting: RTCA Special Committee 226, Audio Systems and Equipment

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-06-25

    ... Federal Aviation Administration Seventh Meeting: RTCA Special Committee 226, Audio Systems and Equipment... Notice of RTCA Special Committee 226, Audio Systems and Equipment. SUMMARY: The FAA is issuing this notice to advise the public of the sixth meeting of the RTCA Special Committee 226, Audio Systems...

  7. 76 FR 79755 - First Meeting: RTCA Special Committee 226 Audio Systems and Equipment

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-12-22

    ... Federal Aviation Administration First Meeting: RTCA Special Committee 226 Audio Systems and Equipment... RTCA Special Committee 226, Audio Systems and Equipment. SUMMARY: The FAA is issuing this notice to advise the public of a meeting of RTCA Special Committee 226, Audio Systems and Equipment, for the...

  8. Design and Usability Testing of an Audio Platform Game for Players with Visual Impairments

    ERIC Educational Resources Information Center

    Oren, Michael; Harding, Chris; Bonebright, Terri L.

    2008-01-01

    This article reports on the evaluation of a novel audio platform game that creates a spatial, interactive experience via audio cues. A pilot study with players with visual impairments, and usability testing comparing the visual and audio game versions using both sighted players and players with visual impairments, revealed that all the…

  9. 77 FR 37732 - Fourteenth Meeting: RTCA Special Committee 224, Audio Systems and Equipment

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-06-22

    ... Federal Aviation Administration Fourteenth Meeting: RTCA Special Committee 224, Audio Systems and...: Meeting Notice of RTCA Special Committee 224, Audio Systems and Equipment. SUMMARY: The FAA is issuing this notice to advise the public of the fourteenth meeting of RTCA Special Committee 224, Audio...

  10. The Use of Asynchronous Audio Feedback with Online RN-BSN Students

    ERIC Educational Resources Information Center

    London, Julie E.

    2013-01-01

    The use of audio technology by online nursing educators is a recent phenomenon. Research has been conducted in the area of audio technology in different domains and populations, but very few researchers have focused on nursing. Preliminary results have indicated that using audio in place of text can increase student cognition and socialization.…

  11. 78 FR 57673 - Eighth Meeting: RTCA Special Committee 226, Audio Systems and Equipment

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-09-19

    ... Federal Aviation Administration Eighth Meeting: RTCA Special Committee 226, Audio Systems and Equipment... Notice of RTCA Special Committee 226, Audio Systems and Equipment. SUMMARY: The FAA is issuing this notice to advise the public of the eighth meeting of the RTCA Special Committee 226, Audio Systems...

  12. Development and Assessment of Web Courses That Use Streaming Audio and Video Technologies.

    ERIC Educational Resources Information Center

    Ingebritsen, Thomas S.; Flickinger, Kathleen

    Iowa State University, through a program called Project BIO (Biology Instructional Outreach), has been using RealAudio technology for about 2 years in college biology courses that are offered entirely via the World Wide Web. RealAudio is a type of streaming media technology that can be used to deliver audio content and a variety of other media…

  13. Comparing observer models and feature selection methods for a task-based statistical assessment of digital breast tomsynthesis in reconstruction space

    NASA Astrophysics Data System (ADS)

    Park, Subok; Zhang, George Z.; Zeng, Rongping; Myers, Kyle J.

    2014-03-01

    A task-based assessment of image quality1 for digital breast tomosynthesis (DBT) can be done in either the projected or reconstructed data space. As the choice of observer models and feature selection methods can vary depending on the type of task and data statistics, we previously investigated the performance of two channelized- Hotelling observer models in conjunction with 2D Laguerre-Gauss (LG) and two implementations of partial least squares (PLS) channels along with that of the Hotelling observer in binary detection tasks involving DBT projections.2, 3 The difference in these observers lies in how the spatial correlation in DBT angular projections is incorporated in the observer's strategy to perform the given task. In the current work, we extend our method to the reconstructed data space of DBT. We investigate how various model observers including the aforementioned compare for performing the binary detection of a spherical signal embedded in structured breast phantoms with the use of DBT slices reconstructed via filtered back projection. We explore how well the model observers incorporate the spatial correlation between different numbers of reconstructed DBT slices while varying the number of projections. For this, relatively small and large scan angles (24° and 96°) are used for comparison. Our results indicate that 1) given a particular scan angle, the number of projections needed to achieve the best performance for each observer is similar across all observer/channel combinations, i.e., Np = 25 for scan angle 96° and Np = 13 for scan angle 24°, and 2) given these sufficient numbers of projections, the number of slices for each observer to achieve the best performance differs depending on the channel/observer types, which is more pronounced in the narrow scan angle case.

  14. Unsupervised Decoding of Long-Term, Naturalistic Human Neural Recordings with Automated Video and Audio Annotations.

    PubMed

    Wang, Nancy X R; Olson, Jared D; Ojemann, Jeffrey G; Rao, Rajesh P N; Brunton, Bingni W

    2016-01-01

    Fully automated decoding of human activities and intentions from direct neural recordings is a tantalizing challenge in brain-computer interfacing. Implementing Brain Computer Interfaces (BCIs) outside carefully controlled experiments in laboratory settings requires adaptive and scalable strategies with minimal supervision. Here we describe an unsupervised approach to decoding neural states from naturalistic human brain recordings. We analyzed continuous, long-term electrocorticography (ECoG) data recorded over many days from the brain of subjects in a hospital room, with simultaneous audio and video recordings. We discovered coherent clusters in high-dimensional ECoG recordings using hierarchical clustering and automatically annotated them using speech and movement labels extracted from audio and video. To our knowledge, this represents the first time techniques from computer vision and speech processing have been used for natural ECoG decoding. Interpretable behaviors were decoded from ECoG data, including moving, speaking and resting; the results were assessed by comparison with manual annotation. Discovered clusters were projected back onto the brain revealing features consistent with known functional areas, opening the door to automated functional brain mapping in natural settings. PMID:27148018

  15. Unsupervised Decoding of Long-Term, Naturalistic Human Neural Recordings with Automated Video and Audio Annotations

    PubMed Central

    Wang, Nancy X. R.; Olson, Jared D.; Ojemann, Jeffrey G.; Rao, Rajesh P. N.; Brunton, Bingni W.

    2016-01-01

    Fully automated decoding of human activities and intentions from direct neural recordings is a tantalizing challenge in brain-computer interfacing. Implementing Brain Computer Interfaces (BCIs) outside carefully controlled experiments in laboratory settings requires adaptive and scalable strategies with minimal supervision. Here we describe an unsupervised approach to decoding neural states from naturalistic human brain recordings. We analyzed continuous, long-term electrocorticography (ECoG) data recorded over many days from the brain of subjects in a hospital room, with simultaneous audio and video recordings. We discovered coherent clusters in high-dimensional ECoG recordings using hierarchical clustering and automatically annotated them using speech and movement labels extracted from audio and video. To our knowledge, this represents the first time techniques from computer vision and speech processing have been used for natural ECoG decoding. Interpretable behaviors were decoded from ECoG data, including moving, speaking and resting; the results were assessed by comparison with manual annotation. Discovered clusters were projected back onto the brain revealing features consistent with known functional areas, opening the door to automated functional brain mapping in natural settings. PMID:27148018

  16. Space Shuttle Wireless Crew Communications

    NASA Technical Reports Server (NTRS)

    Armstrong, R. W.; Doe, R. A.

    1982-01-01

    The design, development, and performance characteristics of the Space Shuttle's Wireless Crew Communications System are discussed. This system allows Space Shuttle crews to interface with the onboard audio distribution system without the need for communications umbilicals, and has been designed through the adaptation of commercially available hardware in order to minimize development time. Testing aboard the Space Shuttle Orbiter Columbia has revealed no failures or design deficiencies.

  17. Transcript of Audio Narrative Portion of: Scandinavian Heritage. A Set of Five Audio-Visual Film Strip/Cassette Presentations.

    ERIC Educational Resources Information Center

    Anderson, Gerald D.; Olson, David B.

    The document presents the transcript of the audio narrative portion of approximately 100 interviews with first and second generation Scandinavian immigrants to the United States. The document is intended for use by secondary school classroom teachers as they develop and implement educational programs related to the Scandinavian heritage in…

  18. Deutsch Durch Audio-Visuelle Methode: An Audio-Lingual-Oral Approach to the Teaching of German.

    ERIC Educational Resources Information Center

    Dickinson Public Schools, ND. Instructional Media Center.

    This teaching guide, designed to accompany Chilton's "Deutsch Durch Audio-Visuelle Methode" for German 1 and 2 in a three-year secondary school program, focuses major attention on the operational plan of the program and a student orientation unit. A section on teaching a unit discusses four phases: (1) presentation, (2) explanation, (3)…

  19. Separation of Audio-Visual Speech Sources: A New Approach Exploiting the Audio-Visual Coherence of Speech Stimuli

    NASA Astrophysics Data System (ADS)

    Sodoyer, David; Schwartz, Jean-Luc; Girin, Laurent; Klinkisch, Jacob; Jutten, Christian

    2002-12-01

    We present a new approach to the source separation problem in the case of multiple speech signals. The method is based on the use of automatic lipreading, the objective is to extract an acoustic speech signal from other acoustic signals by exploiting its coherence with the speaker's lip movements. We consider the case of an additive stationary mixture of decorrelated sources, with no further assumptions on independence or non-Gaussian character. Firstly, we present a theoretical framework showing that it is indeed possible to separate a source when some of its spectral characteristics are provided to the system. Then we address the case of audio-visual sources. We show how, if a statistical model of the joint probability of visual and spectral audio input is learnt to quantify the audio-visual coherence, separation can be achieved by maximizing this probability. Finally, we present a number of separation results on a corpus of vowel-plosive-vowel sequences uttered by a single speaker, embedded in a mixture of other voices. We show that separation can be quite good for mixtures of 2, 3, and 5 sources. These results, while very preliminary, are encouraging, and are discussed in respect to their potential complementarity with traditional pure audio separation or enhancement techniques.

  20. Spatial remapping in the audio-tactile ventriloquism effect: a TMS investigation on the role of the ventral intraparietal area.

    PubMed

    Renzi, Chiara; Bruns, Patrick; Heise, Kirstin-Friederike; Zimerman, Maximo; Feldheim, Jan-Frederik; Hummel, Friedhelm C; Röder, Brigitte

    2013-05-01

    Previous studies have suggested that the putative human homologue of the ventral intraparietal area (hVIP) is crucially involved in the remapping of tactile information into external spatial coordinates and in the realignment of tactile and visual maps. It is unclear, however, whether hVIP is critical for the remapping process during audio-tactile cross-modal spatial interactions. The audio-tactile ventriloquism effect, where the perceived location of a sound is shifted toward the location of a synchronous but spatially disparate tactile stimulus, was used to probe spatial interactions in audio-tactile processing. Eighteen healthy volunteers were asked to report the perceived location of brief auditory stimuli presented from three different locations (left, center, and right). Auditory stimuli were presented either alone (unimodal stimuli) or concurrently to a spatially discrepant tactile stimulus applied to the left or right index finger (bimodal stimuli), with the hands adopting either an uncrossed or a crossed posture. Single pulses of TMS were delivered over the hVIP or a control site (primary somatosensory cortex, SI) 80 msec after trial onset. TMS to the hVIP, compared with the control SI-TMS, interfered with the remapping of touch into external space, suggesting that hVIP is crucially involved in transforming spatial reference frames across audition and touch. PMID:23363411

  1. Solid state audio/speech processor analysis

    NASA Astrophysics Data System (ADS)

    Smith, A. R.; Landell, B. P.; Vensko, G.

    1980-03-01

    The progress made in evaluating the feasibility of applying charge coupled devices (CCD's) as well as microprocessors to reduce the cost and complexity of automatic speech recognition (ASR) systems are discussed. Experiments were conducted to measure the speed versus accuracy tradeoffs of four speed up techniques. The techniques were demonstrated to be worthwhile in an efficient realtime automatic word recognition (AWR) system. Microprocessor architectures were designed to implement the realtime AWR system and then evaluated in terms of cost and complexity using three different microprocessors: the 8 bit Intel 8085A; the 16 bit Motorola MC68000; and a 16 bit configuration of the AMD 2901A. Of the three, the AMD 2901A proved preferable from both a cost and a performance standpoint. Hardware cost projections for an AWR system featuring an AMD 2901A architecture and a bandpass filter CCD analyzer indicate that the hardware components for such a system should range between $1,500 for a 52 word vocabulary, to about $12,700 for a 780 word vocabulary.

  2. Our nation's wetlands (video). Audio-Visual

    SciTech Connect

    Not Available

    1990-01-01

    The Department of the Interior is custodian of approximately 500 million acres of federally owned land and has an important role to play in the management of wetlands. To contribute to the President's goal of no net loss of America's remaining wetlands, the Department of the Interior has initiated a 3-point program consisting of wetlands protection, restoration, and research: Wetlands Protection--Reduce wetlands losses on federally owned lands and encourage state and private landholders to practice wetlands conservation; Wetlands Restoration--Increase wetlands gains through the restoration and creation of wetlands on both public and private lands; Wetlands Research--Provide a foundation of scientific knowledge to guide future actions and decisions about wetlands. The audiovisual is a slide/tape-to-video transfer illustrating the various ways Interior bureaus are working to preserve our Nation's wetlands. The tape features an introduction by Secretary Manuel Lujan on the importance of wetlands and recognizing the benefit of such programs as the North American Waterfowl Management Program.

  3. Evaluation of embedded audio feedback on writing assignments.

    PubMed

    Graves, Janet K; Goodman, Joely T; Hercinger, Maribeth; Minnich, Margo; Murcek, Christina M; Parks, Jane M; Shirley, Nancy

    2015-01-01

    The purpose of this pilot study was to compare embedded audio feedback (EAF), which faculty provided using the iPad(®) application iAnnotate(®) PDF to insert audio comments and written feedback (WF), inserted electronically on student papers in a series of writing assignments. Goals included determining whether EAF provides more useful guidance to students than WF and whether EAF promotes connectedness among students and faculty. An additional goal was to ascertain the efficiency and acceptance of EAF as a grading tool by nursing faculty. The pilot study was a quasi-experimental, cross-over, posttest-only design. The project was completed in an Informatics in Health Care course. Faculty alternated the two feedback methods on four papers written by each student. Results of surveys and focus groups revealed that students and faculty had mixed feelings about this technology. Student preferences were equally divided between EAF and WF, with 35% for each, and 28% were undecided. PMID:25535759

  4. Dynamic range control of audio signals by digital signal processing

    NASA Astrophysics Data System (ADS)

    Gilchrist, N. H. C.

    It is often necessary to reduce the dynamic range of musical programs, particularly those comprising orchestral and choral music, for them to be received satisfactorily by listeners to conventional FM and AM broadcasts. With the arrival of DAB (Digital Audio Broadcasting) a much wider dynamic range will become available for radio broadcasting, although some listeners may prefer to have a signal with a reduced dynamic range. This report describes a digital processor developed by the BBC to control the dynamic range of musical programs in a manner similar to that of a trained Studio Manager. It may be used prior to transmission in conventional broadcasting, replacing limiters or other compression equipment. In DAB, it offers the possibility of providing a dynamic range control signal to be sent to the receiver via an ancillary data channel, simultaneously with the uncompressed audio, giving the listener the option of the full dynamic range or a reduced dynamic range.

  5. Characterization of HF Propagation for Digital Audio Broadcasting

    NASA Technical Reports Server (NTRS)

    Vaisnys, Arvydas

    1997-01-01

    The purpose of this presentation is to give a brief overview of some propagation measurements in the Short Wave (3-30 MHz) bands, made in support of a digital audio transmission system design for the Voice of America. This task is a follow on to the Digital Broadcast Satellite Radio task, during which several mitigation techniques would be applicable to digital audio in the Short Wave bands as well, in spite of the differences in propagation impairments in these two bands. Two series of propagation measurements were made to quantify the range of impairments that could be expected. An assessment of the performance of a prototype version of the receiver was also made.

  6. A calculable, transportable audio-frequency AC reference standard

    SciTech Connect

    Oldham, N.M.; Hetrick, P.S. ); Zeng, X. )

    1989-04-01

    A transportable ac voltage source is described, in which sinusoidal signals are synthesized digitally in the audio-frequency range. The rms value of the output waveform may be calculated by measuring the dc level of the individual steps used to generate the waveform. The uncertainty of this calculation at the 7-V level is typically less than +-5 ppm from 60 Hz to 2 kHz and less than +-10 ppm from 30 Hz to 15 kHz.

  7. Applying Spatial Audio to Human Interfaces: 25 Years of NASA Experience

    NASA Technical Reports Server (NTRS)

    Begault, Durand R.; Wenzel, Elizabeth M.; Godfrey, Martine; Miller, Joel D.; Anderson, Mark R.

    2010-01-01

    From the perspective of human factors engineering, the inclusion of spatial audio within a human-machine interface is advantageous from several perspectives. Demonstrated benefits include the ability to monitor multiple streams of speech and non-speech warning tones using a cocktail party advantage, and for aurally-guided visual search. Other potential benefits include the spatial coordination and interaction of multimodal events, and evaluation of new communication technologies and alerting systems using virtual simulation. Many of these technologies were developed at NASA Ames Research Center, beginning in 1985. This paper reviews examples and describes the advantages of spatial sound in NASA-related technologies, including space operations, aeronautics, and search and rescue. The work has involved hardware and software development as well as basic and applied research.

  8. Social Catalysts for Creating Sociable Media Spaces

    NASA Astrophysics Data System (ADS)

    Karahalios, Karrie G.

    Mediated communication between public spaces is a relatively new concept. One current example of this interaction is video conferencing among people within the same organization. Large-scale video conferencing walls have begun to appear in public or semipublic areas, such as workplace lobbies and kitchens. These connections provide a link via audio and/or video to another public space within the organization.

  9. On the Acoustics of Emotion in Audio: What Speech, Music, and Sound have in Common

    PubMed Central

    Weninger, Felix; Eyben, Florian; Schuller, Björn W.; Mortillaro, Marcello; Scherer, Klaus R.

    2013-01-01

    Without doubt, there is emotional information in almost any kind of sound received by humans every day: be it the affective state of a person transmitted by means of speech; the emotion intended by a composer while writing a musical piece, or conveyed by a musician while performing it; or the affective state connected to an acoustic event occurring in the environment, in the soundtrack of a movie, or in a radio play. In the field of affective computing, there is currently some loosely connected research concerning either of these phenomena, but a holistic computational model of affect in sound is still lacking. In turn, for tomorrow’s pervasive technical systems, including affective companions and robots, it is expected to be highly beneficial to understand the affective dimensions of “the sound that something makes,” in order to evaluate the system’s auditory environment and its own audio output. This article aims at a first step toward a holistic computational model: starting from standard acoustic feature extraction schemes in the domains of speech, music, and sound analysis, we interpret the worth of individual features across these three domains, considering four audio databases with observer annotations in the arousal and valence dimensions. In the results, we find that by selection of appropriate descriptors, cross-domain arousal, and valence regression is feasible achieving significant correlations with the observer annotations of up to 0.78 for arousal (training on sound and testing on enacted speech) and 0.60 for valence (training on enacted speech and testing on music). The high degree of cross-domain consistency in encoding the two main dimensions of affect may be attributable to the co-evolution of speech and music from multimodal affect bursts, including the integration of nature sounds for expressive effects. PMID:23750144

  10. High Capacity Reversible Watermarking for Audio by Histogram Shifting and Predicted Error Expansion

    PubMed Central

    Wang, Fei; Chen, Zuo

    2014-01-01

    Being reversible, the watermarking information embedded in audio signals can be extracted while the original audio data can achieve lossless recovery. Currently, the few reversible audio watermarking algorithms are confronted with following problems: relatively low SNR (signal-to-noise) of embedded audio; a large amount of auxiliary embedded location information; and the absence of accurate capacity control capability. In this paper, we present a novel reversible audio watermarking scheme based on improved prediction error expansion and histogram shifting. First, we use differential evolution algorithm to optimize prediction coefficients and then apply prediction error expansion to output stego data. Second, in order to reduce location map bits length, we introduced histogram shifting scheme. Meanwhile, the prediction error modification threshold according to a given embedding capacity can be computed by our proposed scheme. Experiments show that this algorithm improves the SNR of embedded audio signals and embedding capacity, drastically reduces location map bits length, and enhances capacity control capability. PMID:25097883

  11. AudioSense: Enabling Real-time Evaluation of Hearing Aid Technology In-Situ.

    PubMed

    Hasan, Syed Shabih; Lai, Farley; Chipara, Octav; Wu, Yu-Hsiang

    2013-01-01

    AudioSense integrates mobile phones and web technology to measure hearing aid performance in real-time and in-situ. Measuring the performance of hearing aids in the real world poses significant challenges as it depends on the patient's listening context. AudioSense uses Ecological Momentary Assessment methods to evaluate both the perceived hearing aid performance as well as to characterize the listening environment using electronic surveys. AudioSense further characterizes a patient's listening context by recording their GPS location and sound samples. By creating a time-synchronized record of listening performance and listening contexts, AudioSense will allow researchers to understand the relationship between listening context and hearing aid performance. Performance evaluation shows that AudioSense is reliable, energy-efficient, and can estimate Signal-to-Noise Ratio (SNR) levels from captured audio samples. PMID:25013874

  12. Spatial Audio on the Web: Or Why Can't I hear Anything Over There?

    NASA Technical Reports Server (NTRS)

    Wenzel, Elizabeth M.; Schlickenmaier, Herbert (Technical Monitor); Johnson, Gerald (Technical Monitor); Frey, Mary Anne (Technical Monitor); Schneider, Victor S. (Technical Monitor); Ahunada, Albert J. (Technical Monitor)

    1997-01-01

    Auditory complexity, freedom of movement and interactivity is not always possible in a "true" virtual environment, much less in web-based audio. However, a lot of the perceptual and engineering constraints (and frustrations) that researchers, engineers and listeners have experienced in virtual audio are relevant to spatial audio on the web. My talk will discuss some of these engineering constraints and their perceptual consequences, and attempt to relate these issues to implementation on the web.

  13. Variability in Multisensory Responses Predicts the Self-Space.

    PubMed

    Serino, Andrea

    2016-03-01

    Our brains distinguish between stimuli that are close enough to interact with our bodies and those that are further away by generating a multisensory representation of space near the self, termed peripersonal space. Recent findings show that variability in neuronal response to audio-tactile stimuli predicts the location of the peripersonal space boundary at the individual level. PMID:26833067

  14. ASTP video tape recorder ground support equipment (audio/CTE splitter/interleaver). Operations manual

    NASA Technical Reports Server (NTRS)

    1974-01-01

    A descriptive handbook for the audio/CTE splitter/interleaver (RCA part No. 8673734-502) was presented. This unit is designed to perform two major functions: extract audio and time data from an interleaved video/audio signal (splitter section), and provide a test interleaved video/audio/CTE signal for the system (interleaver section). It is a rack mounting unit 7 inches high, 19 inches wide, 20 inches deep, mounted on slides for retracting from the rack, and weighs approximately 40 pounds. The following information is provided: installation, operation, principles of operation, maintenance, schematics and parts lists.

  15. Realization of guitar audio effects using methods of digital signal processing

    NASA Astrophysics Data System (ADS)

    Buś, Szymon; Jedrzejewski, Konrad

    2015-09-01

    The paper is devoted to studies on possibilities of realization of guitar audio effects by means of methods of digital signal processing. As a result of research, some selected audio effects corresponding to the specifics of guitar sound were realized as the real-time system called Digital Guitar Multi-effect. Before implementation in the system, the selected effects were investigated using the dedicated application with a graphical user interface created in Matlab environment. In the second stage, the real-time system based on a microcontroller and an audio codec was designed and realized. The system is designed to perform audio effects on the output signal of an electric guitar.

  16. Paper-Based Textbooks with Audio Support for Print-Disabled Students.

    PubMed

    Fujiyoshi, Akio; Ohsawa, Akiko; Takaira, Takuya; Tani, Yoshiaki; Fujiyoshi, Mamoru; Ota, Yuko

    2015-01-01

    Utilizing invisible 2-dimensional codes and digital audio players with a 2-dimensional code scanner, we developed paper-based textbooks with audio support for students with print disabilities, called "multimodal textbooks." Multimodal textbooks can be read with the combination of the two modes: "reading printed text" and "listening to the speech of the text from a digital audio player with a 2-dimensional code scanner." Since multimodal textbooks look the same as regular textbooks and the price of a digital audio player is reasonable (about 30 euro), we think multimodal textbooks are suitable for students with print disabilities in ordinary classrooms. PMID:26294447

  17. A Virtual Audio Guidance and Alert System for Commercial Aircraft Operations

    NASA Technical Reports Server (NTRS)

    Begault, Durand R.; Wenzel, Elizabeth M.; Shrum, Richard; Miller, Joel; Null, Cynthia H. (Technical Monitor)

    1996-01-01

    Our work in virtual reality systems at NASA Ames Research Center includes the area of aurally-guided visual search, using specially-designed audio cues and spatial audio processing (also known as virtual or "3-D audio") techniques (Begault, 1994). Previous studies at Ames had revealed that use of 3-D audio for Traffic Collision Avoidance System (TCAS) advisories significantly reduced head-down time, compared to a head-down map display (0.5 sec advantage) or no display at all (2.2 sec advantage) (Begault, 1993, 1995; Begault & Pittman, 1994; see Wenzel, 1994, for an audio demo). Since the crew must keep their head up and looking out the window as much as possible when taxiing under low-visibility conditions, and the potential for "blunder" is increased under such conditions, it was sensible to evaluate the audio spatial cueing for a prototype audio ground collision avoidance warning (GCAW) system, and a 3-D audio guidance system. Results were favorable for GCAW, but not for the audio guidance system.

  18. Informed spectral analysis: audio signal parameter estimation using side information

    NASA Astrophysics Data System (ADS)

    Fourer, Dominique; Marchand, Sylvain

    2013-12-01

    Parametric models are of great interest for representing and manipulating sounds. However, the quality of the resulting signals depends on the precision of the parameters. When the signals are available, these parameters can be estimated, but the presence of noise decreases the resulting precision of the estimation. Furthermore, the Cramér-Rao bound shows the minimal error reachable with the best estimator, which can be insufficient for demanding applications. These limitations can be overcome by using the coding approach which consists in directly transmitting the parameters with the best precision using the minimal bitrate. However, this approach does not take advantage of the information provided by the estimation from the signal and may require a larger bitrate and a loss of compatibility with existing file formats. The purpose of this article is to propose a compromised approach, called the 'informed approach,' which combines analysis with (coded) side information in order to increase the precision of parameter estimation using a lower bitrate than pure coding approaches, the audio signal being known. Thus, the analysis problem is presented in a coder/decoder configuration where the side information is computed and inaudibly embedded into the mixture signal at the coder. At the decoder, the extra information is extracted and is used to assist the analysis process. This study proposes applying this approach to audio spectral analysis using sinusoidal modeling which is a well-known model with practical applications and where theoretical bounds have been calculated. This work aims at uncovering new approaches for audio quality-based applications. It provides a solution for challenging problems like active listening of music, source separation, and realistic sound transformations.

  19. Music and audio - oh how they can stress your network

    NASA Astrophysics Data System (ADS)

    Fletcher, R.

    Nearly ten years ago a paper written by the Audio Engineering Society (AES)[1] made a number of interesting statements: 1. 2. The current Internet is inadequate for transmitting music and professional audio. Performance and collaboration across a distance stress beyond acceptable bounds the quality of service Audio and music provide test cases in which the bounds of the network are quickly reached and through which the defects in a network are readily perceived. Given these key points, where are we now? Have we started to solve any of the problems from the musician's point of view? What is it that musician would like to do that can cause the network so many problems? To understand this we need to appreciate that a trained musician's ears are extremely sensitive to very subtle shifts in temporal materials and localisation information. A shift of a few milliseconds can cause difficulties. So, can modern networks provide the temporal accuracy demanded at this level? The sample and bit rates needed to represent music in the digital domain is still contentious, but a general consensus in the professional world is for 96 KHz and IEEE 64-bit floating point. If this was to be run between two points on the network across 24 channels in near real time to allow for collaborative composition/production/performance, with QOS settings to allow as near to zero latency and jitter, it can be seen that the network indeed has to perform very well. Lighting the Blue Touchpaper for UK e-Science - Closing Conference of ESLEA Project The George Hotel, Edinburgh, UK 26-28 March, 200

  20. Fast Huffman encoding algorithms in MPEG-4 advanced audio coding

    NASA Astrophysics Data System (ADS)

    Brzuchalski, Grzegorz

    2014-11-01

    This paper addresses the optimisation problem of Huffman encoding in MPEG-4 Advanced Audio Coding stan- dard. At first, the Huffman encoding problem and the need of encoding two side info parameters scale factor and Huffman codebook are presented. Next, Two Loop Search, Maximum Noise Mask Ratio and Trellis Based algorithms of bit allocation are briefly described. Further, Huffman encoding optimisation are shown. New methods try to check and change scale factor bands as little as possible to estimate bitrate cost or its change. Finally, the complexity of old and new methods is calculated, compared and measured time of encoding is given.

  1. [Voix d'Or, an audio tool to revive memories].

    PubMed

    Braunschweig, Lina

    2010-01-01

    Voix d'Or is an audio tool designed to awaken the affective memory of elderly people and particularly those suffering from Alzheimer's disease. Every month it offers new radio programmes to initiate or facilitate leisure and entertainment activities, memory workshops or provide the basis of quiet moments. The tool has a double objective: to procure well-being, boost the individual's self-esteem and recognise his/her history and to facilitate exchange and communication between the residents and the staff of a care home. PMID:21137493

  2. Audio-vocal responses elicited in adult cochlear implant users

    PubMed Central

    Loucks, Torrey M.; Suneel, Deepa; Aronoff, Justin M.

    2015-01-01

    Auditory deprivation experienced prior to receiving a cochlear implant could compromise neural connections that allow for modulation of vocalization using auditory feedback. In this report, pitch-shift stimuli were presented to adult cochlear implant users to test whether compensatory motor changes in vocal F0 could be elicited. In five of six participants, rapid adjustments in vocal F0 were detected following the stimuli, which resemble the cortically mediated pitch-shift responses observed in typical hearing individuals. These findings suggest that cochlear implants can convey vocal F0 shifts to the auditory pathway that might benefit audio-vocal monitoring. PMID:26520350

  3. TV audio and video on the same channel

    NASA Technical Reports Server (NTRS)

    Hopkins, J. B.

    1979-01-01

    Transmitting technique adds audio to video signal during vertical blanking interval. SIVI (signal in the vertical interval) is used by TV networks and stations to transmit cuing and automatic-switching tone signals to augment automatic and manual operations. It can also be used to transmit one-way instructional information, such as bulletin alerts, program changes, and commercial-cutaway aural cues from the networks to affiliates. Additonally, it can be used as extra sound channel for second-language transmission to biligual stations.

  4. Quantitative characterisation of audio data by ordinal symbolic dynamics

    NASA Astrophysics Data System (ADS)

    Aschenbrenner, T.; Monetti, R.; Amigó, J. M.; Bunk, W.

    2013-06-01

    Ordinal symbolic dynamics has developed into a valuable method to describe complex systems. Recently, using the concept of transcripts, the coupling behaviour of systems was assessed, combining the properties of the symmetric group with information theoretic ideas. In this contribution, methods from the field of ordinal symbolic dynamics are applied to the characterisation of audio data. Coupling complexity between frequency bands of solo violin music, as a fingerprint of the instrument, is used for classification purposes within a support vector machine scheme. Our results suggest that coupling complexity is able to capture essential characteristics, sufficient to distinguish among different violins.

  5. Reducing audio stimulus presentation latencies across studies, laboratories, and hardware and operating system configurations.

    PubMed

    Babjack, Destiny L; Cernicky, Brandon; Sobotka, Andrew J; Basler, Lee; Struthers, Devon; Kisic, Richard; Barone, Kimberly; Zuccolotto, Anthony P

    2015-09-01

    Using differing computer platforms and audio output devices to deliver audio stimuli often introduces (1) substantial variability across labs and (2) variable time between the intended and actual sound delivery (the sound onset latency). Fast, accurate audio onset latencies are particularly important when audio stimuli need to be delivered precisely as part of studies that depend on accurate timing (e.g., electroencephalographic, event-related potential, or multimodal studies), or in multisite studies in which standardization and strict control over the computer platforms used is not feasible. This research describes the variability introduced by using differing configurations and introduces a novel approach to minimizing audio sound latency and variability. A stimulus presentation and latency assessment approach is presented using E-Prime and Chronos (a new multifunction, USB-based data presentation and collection device). The present approach reliably delivers audio stimuli with low latencies that vary by ≤1 ms, independent of hardware and Windows operating system (OS)/driver combinations. The Chronos audio subsystem adopts a buffering, aborting, querying, and remixing approach to the delivery of audio, to achieve a consistent 1-ms sound onset latency for single-sound delivery, and precise delivery of multiple sounds that achieves standard deviations of 1/10th of a millisecond without the use of advanced scripting. Chronos's sound onset latencies are small, reliable, and consistent across systems. Testing of standard audio delivery devices and configurations highlights the need for careful attention to consistency between labs, experiments, and multiple study sites in their hardware choices, OS selections, and adoption of audio delivery systems designed to sidestep the audio latency variability issue. PMID:26170050

  6. A compact electroencephalogram recording device with integrated audio stimulation system.

    PubMed

    Paukkunen, Antti K O; Kurttio, Anttu A; Leminen, Miika M; Sepponen, Raimo E

    2010-06-01

    A compact (96 x 128 x 32 mm(3), 374 g), battery-powered, eight-channel electroencephalogram recording device with an integrated audio stimulation system and a wireless interface is presented. The recording device is capable of producing high-quality data, while the operating time is also reasonable for evoked potential studies. The effective measurement resolution is about 4 nV at 200 Hz sample rate, typical noise level is below 0.7 microV(rms) at 0.16-70 Hz, and the estimated operating time is 1.5 h. An embedded audio decoder circuit reads and plays wave sound files stored on a memory card. The activities are controlled by an 8 bit main control unit which allows accurate timing of the stimuli. The interstimulus interval jitter measured is less than 1 ms. Wireless communication is made through bluetooth and the data recorded are transmitted to an external personal computer (PC) interface in real time. The PC interface is implemented with LABVIEW and in addition to data acquisition it also allows online signal processing, data storage, and control of measurement activities such as contact impedance measurement, for example. The practical application of the device is demonstrated in mismatch negativity experiment with three test subjects. PMID:20590254

  7. Audio watermarking technologies for automatic cue sheet generation systems

    NASA Astrophysics Data System (ADS)

    Caccia, Giuseppe; Lancini, Rosa C.; Pascarella, Annalisa; Tubaro, Stefano; Vicario, Elena

    2001-08-01

    Usually watermark is used as a way for hiding information on digital media. The watermarked information may be used to allow copyright protection or user and media identification. In this paper we propose a watermarking scheme for digital audio signals that allow automatic identification of musical pieces transmitted in TV broadcasting programs. In our application the watermark must be, obviously, imperceptible to the users, should be robust to standard TV and radio editing and have a very low complexity. This last item is essential to allow a software real-time implementation of the insertion and detection of watermarks using only a minimum amount of the computation power of a modern PC. In the proposed method the input audio sequence is subdivided in frames. For each frame a watermark spread spectrum sequence is added to the original data. A two steps filtering procedure is used to generate the watermark from a Pseudo-Noise (PN) sequence. The filters approximate respectively the threshold and the frequency masking of the Human Auditory System (HAS). In the paper we discuss first the watermark embedding system then the detection approach. The results of a large set of subjective tests are also presented to demonstrate the quality and robustness of the proposed approach.

  8. Information-Driven Active Audio-Visual Source Localization

    PubMed Central

    Schult, Niclas; Reineking, Thomas; Kluss, Thorsten; Zetzsche, Christoph

    2015-01-01

    We present a system for sensorimotor audio-visual source localization on a mobile robot. We utilize a particle filter for the combination of audio-visual information and for the temporal integration of consecutive measurements. Although the system only measures the current direction of the source, the position of the source can be estimated because the robot is able to move and can therefore obtain measurements from different directions. These actions by the robot successively reduce uncertainty about the source’s position. An information gain mechanism is used for selecting the most informative actions in order to minimize the number of actions required to achieve accurate and precise position estimates in azimuth and distance. We show that this mechanism is an efficient solution to the action selection problem for source localization, and that it is able to produce precise position estimates despite simplified unisensory preprocessing. Because of the robot’s mobility, this approach is suitable for use in complex and cluttered environments. We present qualitative and quantitative results of the system’s performance and discuss possible areas of application. PMID:26327619

  9. Audio-Tactile Integration and the Influence of Musical Training

    PubMed Central

    Kuchenbuch, Anja; Paraskevopoulos, Evangelos; Herholz, Sibylle C.; Pantev, Christo

    2014-01-01

    Perception of our environment is a multisensory experience; information from different sensory systems like the auditory, visual and tactile is constantly integrated. Complex tasks that require high temporal and spatial precision of multisensory integration put strong demands on the underlying networks but it is largely unknown how task experience shapes multisensory processing. Long-term musical training is an excellent model for brain plasticity because it shapes the human brain at functional and structural levels, affecting a network of brain areas. In the present study we used magnetoencephalography (MEG) to investigate how audio-tactile perception is integrated in the human brain and if musicians show enhancement of the corresponding activation compared to non-musicians. Using a paradigm that allowed the investigation of combined and separate auditory and tactile processing, we found a multisensory incongruency response, generated in frontal, cingulate and cerebellar regions, an auditory mismatch response generated mainly in the auditory cortex and a tactile mismatch response generated in frontal and cerebellar regions. The influence of musical training was seen in the audio-tactile as well as in the auditory condition, indicating enhanced higher-order processing in musicians, while the sources of the tactile MMN were not influenced by long-term musical training. Consistent with the predictive coding model, more basic, bottom-up sensory processing was relatively stable and less affected by expertise, whereas areas for top-down models of multisensory expectancies were modulated by training. PMID:24465675

  10. Separate Mechanisms for Audio-Tactile Pitch and Loudness Interactions

    PubMed Central

    Yau, Jeffrey M.; Weber, Alison I.; Bensmaia, Sliman J.

    2010-01-01

    A major goal in perceptual neuroscience is to understand how signals from different sensory modalities are combined to produce stable and coherent representations. We previously investigated interactions between audition and touch, motivated by the fact that both modalities are sensitive to environmental oscillations. In our earlier study, we characterized the effect of auditory distractors on tactile frequency and intensity perception. Here, we describe the converse experiments examining the effect of tactile distractors on auditory processing. Because the two studies employ the same psychophysical paradigm, we combined their results for a comprehensive view of how auditory and tactile signals interact and how these interactions depend on the perceptual task. Together, our results show that temporal frequency representations are perceptually linked regardless of the attended modality. In contrast, audio-tactile loudness interactions depend on the attended modality: Tactile distractors influence judgments of auditory intensity, but judgments of tactile intensity are impervious to auditory distraction. Lastly, we show that audio-tactile loudness interactions depend critically on stimulus timing, while pitch interactions do not. These results reveal that auditory and tactile inputs are combined differently depending on the perceptual task. That distinct rules govern the integration of auditory and tactile signals in pitch and loudness perception implies that the two are mediated by separate neural mechanisms. These findings underscore the complexity and specificity of multisensory interactions. PMID:21887147

  11. Characteristics of the audio sound generated by ultrasound imaging systems

    NASA Astrophysics Data System (ADS)

    Fatemi, Mostafa; Alizad, Azra; Greenleaf, James F.

    2005-03-01

    Medical ultrasound scanners use high-energy pulses to probe the human body. The radiation force resulting from the impact of such pulses on an object can vibrate the object, producing a localized high-intensity sound in the audible range. Here, a theoretical model for the audio sound generated by ultrasound scanners is presented. This model describes the temporal and spectral characteristics of the sound. It has been shown that the sound has rich frequency components at the pulse repetition frequency and its harmonics. Experiments have been conducted in a water tank to measure the sound generated by a clinical ultrasound scanner in various operational modes. Results are in general agreement with the theory. It is shown that a typical ultrasound scanner with a typical spatial-peak pulse-average intensity value at 2 MHz may generate a localized sound-pressure level close to 100 dB relative to 20 μPa in the audible (<20 kHz) range under laboratory conditions. These findings suggest that fetuses may become exposed to a high-intensity audio sound during maternal ultrasound examinations. Therefore, contrary to common beliefs, ultrasound may not be considered a passive tool in fetal imaging..

  12. Audio annotation watermarking with robustness against DA/AD conversion

    NASA Astrophysics Data System (ADS)

    Qian, Kun; Kraetzer, Christian; Biermann, Michael; Dittmann, Jana

    2010-01-01

    In the paper we present a watermarking scheme developed to meet the specific requirements of audio annotation watermarking robust against DA/AD conversion (watermark detection after playback by loudspeaker and recording with a microphone). Additionally the described approach tries to achieve a comparably low detection complexity, so it could be embedded in the near future in low-end devices (e.g. mobile phones or other portable devices). We assume in the field of annotation watermarking that there is no specific motivation for attackers to the developed scheme. The basic idea for the watermark generation and embedding scheme is to combine traditional frequency domain spread spectrum watermarking with psychoacoustic modeling to guarantee transparency and alphabet substitution to improve the robustness. The synchronization and extraction scheme is designed to be much less computational complex than the embedder. The performance of the scheme is evaluated in the aspects of transparency, robustness, complexity and capacity. The tests reveals that 44% out of 375 tested audio files pass the simulation test for robustness, while the most appropriate category shows even 100% robustness. Additionally the introduced prototype shows an averge transparency of -1.69 in SDG, while at the same time having a capacity satisfactory to the chosen application scenario.

  13. Information-Driven Active Audio-Visual Source Localization.

    PubMed

    Schult, Niclas; Reineking, Thomas; Kluss, Thorsten; Zetzsche, Christoph

    2015-01-01

    We present a system for sensorimotor audio-visual source localization on a mobile robot. We utilize a particle filter for the combination of audio-visual information and for the temporal integration of consecutive measurements. Although the system only measures the current direction of the source, the position of the source can be estimated because the robot is able to move and can therefore obtain measurements from different directions. These actions by the robot successively reduce uncertainty about the source's position. An information gain mechanism is used for selecting the most informative actions in order to minimize the number of actions required to achieve accurate and precise position estimates in azimuth and distance. We show that this mechanism is an efficient solution to the action selection problem for source localization, and that it is able to produce precise position estimates despite simplified unisensory preprocessing. Because of the robot's mobility, this approach is suitable for use in complex and cluttered environments. We present qualitative and quantitative results of the system's performance and discuss possible areas of application. PMID:26327619

  14. Audio-tactile integration and the influence of musical training.

    PubMed

    Kuchenbuch, Anja; Paraskevopoulos, Evangelos; Herholz, Sibylle C; Pantev, Christo

    2014-01-01

    Perception of our environment is a multisensory experience; information from different sensory systems like the auditory, visual and tactile is constantly integrated. Complex tasks that require high temporal and spatial precision of multisensory integration put strong demands on the underlying networks but it is largely unknown how task experience shapes multisensory processing. Long-term musical training is an excellent model for brain plasticity because it shapes the human brain at functional and structural levels, affecting a network of brain areas. In the present study we used magnetoencephalography (MEG) to investigate how audio-tactile perception is integrated in the human brain and if musicians show enhancement of the corresponding activation compared to non-musicians. Using a paradigm that allowed the investigation of combined and separate auditory and tactile processing, we found a multisensory incongruency response, generated in frontal, cingulate and cerebellar regions, an auditory mismatch response generated mainly in the auditory cortex and a tactile mismatch response generated in frontal and cerebellar regions. The influence of musical training was seen in the audio-tactile as well as in the auditory condition, indicating enhanced higher-order processing in musicians, while the sources of the tactile MMN were not influenced by long-term musical training. Consistent with the predictive coding model, more basic, bottom-up sensory processing was relatively stable and less affected by expertise, whereas areas for top-down models of multisensory expectancies were modulated by training. PMID:24465675

  15. Development and Evaluation of a Feedback Support System with Audio and Playback Strokes

    ERIC Educational Resources Information Center

    Li, Kai; Akahori, Kanji

    2008-01-01

    This paper describes the development and evaluation of a handwritten correction support system with audio and playback strokes used to teach Japanese writing. The study examined whether audio and playback strokes have a positive effect on students using honorific expressions in Japanese writing. The results showed that error feedback with audio…

  16. 47 CFR 73.4275 - Tone clusters; audio attention-getting devices.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 47 Telecommunication 4 2012-10-01 2012-10-01 false Tone clusters; audio attention-getting devices. 73.4275 Section 73.4275 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST... clusters; audio attention-getting devices. See Public Notice, FCC 76-610, dated July 2, 1976. 60 FCC 2d...

  17. 47 CFR 73.4275 - Tone clusters; audio attention-getting devices.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 47 Telecommunication 4 2013-10-01 2013-10-01 false Tone clusters; audio attention-getting devices. 73.4275 Section 73.4275 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST... clusters; audio attention-getting devices. See Public Notice, FCC 76-610, dated July 2, 1976. 60 FCC 2d...

  18. 47 CFR 73.4275 - Tone clusters; audio attention-getting devices.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 47 Telecommunication 4 2011-10-01 2011-10-01 false Tone clusters; audio attention-getting devices. 73.4275 Section 73.4275 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST... clusters; audio attention-getting devices. See Public Notice, FCC 76-610, dated July 2, 1976. 60 FCC 2d...

  19. 47 CFR 73.4275 - Tone clusters; audio attention-getting devices.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 47 Telecommunication 4 2014-10-01 2014-10-01 false Tone clusters; audio attention-getting devices. 73.4275 Section 73.4275 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST... clusters; audio attention-getting devices. See Public Notice, FCC 76-610, dated July 2, 1976. 60 FCC 2d...

  20. Age Matters: Student Experiences with Audio Learning Guides in University-Based Continuing Education

    ERIC Educational Resources Information Center

    Mercer, Lorraine; Pianosi, Birgit

    2012-01-01

    The primary objective of this research was to explore the experiences of undergraduate distance education students using sample audio versions (provided on compact disc) of the learning guides for their courses. The results of this study indicated that students responded positively to the opportunity to have word-for-word audio versions of their…

  1. Reaching Out: The Role of Audio Cassette Communication in Rural Development. Occasional Paper 19.

    ERIC Educational Resources Information Center

    Adhikarya, Ronny; Colle, Royal D.

    This report describes the state-of-the-art of audio cassette technology (ACT) and reports findings from field tests, case studies, and pilot projects in several countries which demonstrate the potential of audio cassettes as a medium for communicating with rural people. Specific guidance is also offered on how a project can use cassettes as a…

  2. LiveDescribe: Can Amateur Describers Create High-Quality Audio Description?

    ERIC Educational Resources Information Center

    Branje, Carmen J.; Fels, Deborah I.

    2012-01-01

    Introduction: The study presented here evaluated the usability of the audio description software LiveDescribe and explored the acceptance rates of audio description created by amateur describers who used LiveDescribe to facilitate the creation of their descriptions. Methods: Twelve amateur describers with little or no previous experience with…

  3. 36 CFR 5.5 - Commercial filming, still photography, and audio recording.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... of 43 CFR part 5, subpart A. Failure to comply with any provision of 43 CFR part 5 is a violation of... schedule for still photography conducted under a permit issued under 43 CFR part 5 applies to audio... photography, and audio recording. 5.5 Section 5.5 Parks, Forests, and Public Property NATIONAL PARK...

  4. Rethinking the Red Ink: Audio-Feedback in the ESL Writing Classroom.

    ERIC Educational Resources Information Center

    Johanson, Robert

    1999-01-01

    This paper describes audio-feedback as a teaching method for English-as-a-Second-Language (ESL) writing classes. Using this method, writing instructors respond to students' compositions by recording their comments onto an audiocassette, then returning the paper and cassette to the students. The first section describes audio-feedback and explains…

  5. An Analysis of Certain Elements of an Audio-Tape Approach to Instruction.

    ERIC Educational Resources Information Center

    Bell, Ronald Ernest

    This study was designed to determine the association between selected variables and an audio-tape approach to instruction. Fifty sophomore students enrolled in a physical anthropology course at Shoreline Community College (Washington) participated in an experimental instructional program that consisted of thirty-two audio-tapes and three optional…

  6. Audio-Based versus Text-Based Asynchronous Online Discussion: Two Case Studies

    ERIC Educational Resources Information Center

    Hew, Khe Foon; Cheung, Wing Sum

    2013-01-01

    The main objective of this paper is to examine the use of audio- versus text-based asynchronous online discussions. We report two case studies conducted within the context of semester-long teacher education courses at an Asian Pacific university. Forty-one graduate students participated in Study I. After the online discussions (both audio-based as…

  7. The Language System of Audio Description: An Investigation as a Discursive Process

    ERIC Educational Resources Information Center

    Piety, Philip J.

    2004-01-01

    This study investigated the language used in a selection of films containing audio description and developed a set of definitions that allow productions containing it to be more fully defined, measured, and compared. It also highlights some challenging questions related to audio description as a discursive practice and provides a basis for future…

  8. 16 CFR 307.8 - Requirements for disclosure in audiovisual and audio advertising.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ... in a manner that is clearly audible. If an advertisement has both a visual and an audio component... and audio advertising. 307.8 Section 307.8 Commercial Practices FEDERAL TRADE COMMISSION REGULATIONS... ACT OF 1986 Advertising Disclosures § 307.8 Requirements for disclosure in audiovisual and...

  9. Immediate Audio and Visual Confirmation; "Breakthrough" for the Low-Aptitude Language Student.

    ERIC Educational Resources Information Center

    Mueller, Theodore H.

    Students with low language aptitude have been found to have poor powers of auditory discrimination. To date, programed language instruction has relied on audio confirmation of oral response. A study was conducted to determine the value of adding visual confirmation to the audio model. A total of 170 experimental and 140 control students in second…

  10. 47 CFR 73.9005 - Compliance requirements for covered demodulator products: Audio.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... unscreened content or of marked content in digital form except in compressed audio format (such as AC3) or in... products: Audio. 73.9005 Section 73.9005 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST RADIO SERVICES RADIO BROADCAST SERVICES Digital Broadcast Television Redistribution Control §...

  11. Using TTS Voices to Develop Audio Materials for Listening Comprehension: A Digital Approach

    ERIC Educational Resources Information Center

    Sha, Guoquan

    2010-01-01

    This paper reports a series of experiments with text-to-speech (TTS) voices. These experiments have been conducted to develop audio materials for listening comprehension as an alternative technology to traditionally used audio equipment like the compact cassette. The new generation of TTS voices based on unit selection synthesis provides…

  12. Report to the Legislature: Audio-Digital MCAS Pilot Program. Line Item 7061-0012

    ERIC Educational Resources Information Center

    Massachusetts Department of Elementary and Secondary Education, 2008

    2008-01-01

    This paper presents the Final Report on the Audio-Digital MCAS Pilot Program. The Department and Recording For the Blind & Dyslexic (RFB&D) have collaborated to provide audio-digital read-aloud editions of the Grade 10 English Language Arts and Mathematics MCAS tests for a small number of students with disabilities such as dyslexia and/or vision…

  13. 37 CFR 201.28 - Statements of Account for digital audio recording devices or media.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... 37 Patents, Trademarks, and Copyrights 1 2011-07-01 2011-07-01 false Statements of Account for digital audio recording devices or media. 201.28 Section 201.28 Patents, Trademarks, and Copyrights COPYRIGHT OFFICE, LIBRARY OF CONGRESS COPYRIGHT OFFICE AND PROCEDURES GENERAL PROVISIONS § 201.28 Statements of Account for digital audio...

  14. 77 FR 16890 - Second Meeting: RTCA Special Committee 226, Audio Systems and Equipment

    Federal Register 2010, 2011, 2012, 2013, 2014

    2012-03-22

    ... Federal Aviation Administration Second Meeting: RTCA Special Committee 226, Audio Systems and Equipment... to advise the public of the second meeting of RTCA Special Committee 226, Audio Systems and Equipment.../changes Draft language for noise test requirement to combine with the vibration test Other...

  15. 47 CFR 73.4275 - Tone clusters; audio attention-getting devices.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 47 Telecommunication 4 2010-10-01 2010-10-01 false Tone clusters; audio attention-getting devices. 73.4275 Section 73.4275 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST... clusters; audio attention-getting devices. See Public Notice, FCC 76-610, dated July 2, 1976. 60 FCC 2d...

  16. Effects of Audio-Visual Information on the Intelligibility of Alaryngeal Speech

    ERIC Educational Resources Information Center

    Evitts, Paul M.; Portugal, Lindsay; Van Dine, Ami; Holler, Aline

    2010-01-01

    Background: There is minimal research on the contribution of visual information on speech intelligibility for individuals with a laryngectomy (IWL). Aims: The purpose of this project was to determine the effects of mode of presentation (audio-only, audio-visual) on alaryngeal speech intelligibility. Method: Twenty-three naive listeners were…

  17. When I Stopped Writing on Their Papers: Accommodating the Needs of Student Writers with Audio Comments

    ERIC Educational Resources Information Center

    Bauer, Sara

    2011-01-01

    The author finds using software to make audio comments on students' writing improves students' understanding of her responses and increases their willingness to take her suggestions for revision more seriously. In the process of recording audio comments, she came to a new understanding of her students' writing needs and her responsibilities as…

  18. "Listen to This!" Utilizing Audio Recordings to Improve Instructor Feedback on Writing in Mathematics

    ERIC Educational Resources Information Center

    Weld, Christopher

    2014-01-01

    Providing audio files in lieu of written remarks on graded assignments is arguably a more effective means of feedback, allowing students to better process and understand the critique and improve their future work. With emerging technologies and software, this audio feedback alternative to the traditional paradigm of providing written comments…

  19. Students' Attitudes to and Usage of Academic Feedback Provided via Audio Files

    ERIC Educational Resources Information Center

    Merry, Stephen; Orsmond, Paul

    2008-01-01

    This study explores students' attitudes to the provision of formative feedback on academic work using audio files together with the ways in which students implement such feedback within their learning. Fifteen students received audio file feedback on written work and were subsequently interviewed regarding their utilisation of that feedback within…

  20. Active Learning in the Online Environment: The Integration of Student-Generated Audio Files

    ERIC Educational Resources Information Center

    Bolliger, Doris U.; Armier, David Des, Jr.

    2013-01-01

    Educators have integrated instructor-produced audio files in a variety of settings and environments for purposes such as content presentation, lecture reviews, student feedback, and so forth. Few instructors, however, require students to produce audio files and share them with peers. The purpose of this study was to obtain empirical data on…

  1. Investigating Expectations and Experiences of Audio and Written Assignment Feedback in First-Year Undergraduate Students

    ERIC Educational Resources Information Center

    Fawcett, Hannah; Oldfield, Jeremy

    2016-01-01

    Previous research suggests that audio feedback may be an important mechanism for facilitating effective and timely assignment feedback. The present study examined expectations and experiences of audio and written feedback provided through "turnitin for iPad®" from students within the same cohort and assignment. The results showed that…

  2. Temporal Interval Discrimination Thresholds Depend on Perceived Synchrony for Audio-Visual Stimulus Pairs

    ERIC Educational Resources Information Center

    van Eijk, Rob L. J.; Kohlrausch, Armin; Juola, James F.; van de Par, Steven

    2009-01-01

    Audio-visual stimulus pairs presented at various relative delays, are commonly judged as being "synchronous" over a range of delays from about -50 ms (audio leading) to +150 ms (video leading). The center of this range is an estimate of the point of subjective simultaneity (PSS). The judgment boundaries, where "synchronous" judgments yield to a…

  3. Audio Control Handbook For Radio and Television Broadcasting. Third Revised Edition.

    ERIC Educational Resources Information Center

    Oringel, Robert S.

    Audio control is the operation of all the types of sound equipment found in the studios and control rooms of a radio or television station. Written in a nontechnical style for beginners, the book explains thoroughly the operation of all types of audio equipment. Diagrams and photographs of commercial consoles, microphones, turntables, and tape…

  4. Guidelines for the Production of Audio Materials for Print Handicapped Readers.

    ERIC Educational Resources Information Center

    National Library of Australia, Canberra.

    Procedural guidelines developed by the Audio Standards Committee of the National Library of Australia to help improve the overall quality of production of audio materials for visually handicapped readers are presented. This report covers the following areas: selection of narrators and the narration itself; copyright; recording of books, magazines,…

  5. Changes of the Prefrontal EEG (Electroencephalogram) Activities According to the Repetition of Audio-Visual Learning.

    ERIC Educational Resources Information Center

    Kim, Yong-Jin; Chang, Nam-Kee

    2001-01-01

    Investigates the changes of neuronal response according to a four time repetition of audio-visual learning. Obtains EEG data from the prefrontal (Fp1, Fp2) lobe from 20 subjects at the 8th grade level. Concludes that the habituation of neuronal response shows up in repetitive audio-visual learning and brain hemisphericity can be changed by…

  6. 47 CFR Figure 2 to Subpart N of... - Typical Audio Wave

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 47 Telecommunication 1 2010-10-01 2010-10-01 false Typical Audio Wave 2 Figure 2 to Subpart N of Part 2 Telecommunication FEDERAL COMMUNICATIONS COMMISSION GENERAL FREQUENCY ALLOCATIONS AND RADIO... Audio Wave EC03JN91.006...

  7. Planning Schools for Use of Audio-Visual Materials. No. 1--Classrooms, 3rd Edition.

    ERIC Educational Resources Information Center

    National Education Association, Washington, DC.

    Intended to inform school board administrators and teachers of the current (1958) thinking on audio-visual instruction for use in planning new buildings, purchasing equipment, and planning instruction. Attention is given the problem of overcoming obstacles to the incorporation of audio-visual materials into the curriculum. Discussion includes--(1)…

  8. An Interactive Concert Program Based on Infrared Watermark and Audio Synthesis

    NASA Astrophysics Data System (ADS)

    Wang, Hsi-Chun; Lee, Wen-Pin Hope; Liang, Feng-Ju

    The objective of this research is to propose a video/audio system which allows the user to listen the typical music notes in the concert program under infrared detection. The system synthesizes audio with different pitches and tempi in accordance with the encoded data in a 2-D barcode embedded in the infrared watermark. The digital halftoning technique has been used to fabricate the infrared watermark composed of halftone dots by both amplitude modulation (AM) and frequency modulation (FM). The results show that this interactive system successfully recognizes the barcode and synthesizes audio under infrared detection of a concert program which is also valid for human observation of the contents. This interactive video/audio system has greatly expanded the capability of the printout paper to audio display and also has many potential value-added applications.

  9. Acoustical characteristics of water sounds for soundscape enhancement in urban open spaces.

    PubMed

    Jeon, Jin Yong; Lee, Pyoung Jik; You, Jin; Kang, Jian

    2012-03-01

    The goal of the present study is to characterize water sounds that can be used in urban open spaces to mask road traffic noise. Sounds and visual images of a number of water features located in urban open places were obtained and subsequently analyzed in terms of psychoacoustical metrics and acoustical measures. Laboratory experiments were then conducted to investigate which water sound is appropriate for masking urban noise. The experiments consisted of two sessions: (1) Audio-only condition and (2) combined audio-visual condition. Subjective responses to stimuli were rated through the use of preference scores and 15 adjectives. The results of the experiments revealed that preference scores for the urban soundscape were affected by the acoustical characteristics of water sounds and visual images of water features; Sharpness that was used to explain the spectral envelopes of water sounds was proved to be a dominant factor for urban soundscape perception; and preferences regarding the urban soundscape were significantly related to adjectives describing "freshness" and "calmness." PMID:22423706

  10. Effect of Audio Coaching on Correlation of Abdominal Displacement With Lung Tumor Motion

    SciTech Connect

    Nakamura, Mitsuhiro Narita, Yuichiro; Matsuo, Yukinori; Narabayashi, Masaru; Nakata, Manabu; Sawada, Akira; Mizowaki, Takashi; Nagata, Yasushi; Hiraoka, Masahiro

    2009-10-01

    Purpose: To assess the effect of audio coaching on the time-dependent behavior of the correlation between abdominal motion and lung tumor motion and the corresponding lung tumor position mismatches. Methods and Materials: Six patients who had a lung tumor with a motion range >8 mm were enrolled in the present study. Breathing-synchronized fluoroscopy was performed initially without audio coaching, followed by fluoroscopy with recorded audio coaching for multiple days. Two different measurements, anteroposterior abdominal displacement using the real-time positioning management system and superoinferior (SI) lung tumor motion by X-ray fluoroscopy, were performed simultaneously. Their sequential images were recorded using one display system. The lung tumor position was automatically detected with a template matching technique. The relationship between the abdominal and lung tumor motion was analyzed with and without audio coaching. Results: The mean SI tumor displacement was 10.4 mm without audio coaching and increased to 23.0 mm with audio coaching (p < .01). The correlation coefficients ranged from 0.89 to 0.97 with free breathing. Applying audio coaching, the correlation coefficients improved significantly (range, 0.93-0.99; p < .01), and the SI lung tumor position mismatches became larger in 75% of all sessions. Conclusion: Audio coaching served to increase the degree of correlation and make it more reproducible. In addition, the phase shifts between tumor motion and abdominal displacement were improved; however, all patients breathed more deeply, and the SI lung tumor position mismatches became slightly larger with audio coaching than without audio coaching.

  11. The Brussels Mood Inductive Audio Stories (MIAS) database.

    PubMed

    Bertels, Julie; Deliens, Gaétane; Peigneux, Philippe; Destrebecqz, Arnaud

    2014-12-01

    Through this study, we aimed to validate a new tool for inducing moods in experimental contexts. Five audio stories with sad, joyful, frightening, erotic, or neutral content were presented to 60 participants (33 women, 27 men) in a within-subjects design, each for about 10 min. Participants were asked (1) to report their moods before and after listening to each story, (2) to assess the emotional content of the excerpts on various emotional scales, and (3) to rate their level of projection into the stories. The results confirmed our a priori emotional classification. The emotional stories were effective in inducing the desired mood, with no difference found between male and female participants. These stories therefore constitute a valuable corpus for inducing moods in French-speaking participants, and they are made freely available for use in scientific research. PMID:24519495

  12. Increasing observer objectivity with audio-visual technology: the Sphygmocorder.

    PubMed

    Atkins; O'Brien; Wesseling; Guelen

    1997-10-01

    The most fallible component of blood pressure measurement is the human observer. The traditional technique of measuring blood pressure does not allow the result of the measurement to be checked by independent observers, thereby leaving the method open to bias. In the Sphygmocorder, several components used to measure blood pressure have been combined innovatively with audio-visual recording technology to produce a system consisting of a mercury sphygmomanometer, an occluding cuff, an automatic inflation-deflation source, a stethoscope, a microphone capable of detecting Korotkoff sounds, a camcorder and a display screen. The accuracy of the Sphygmocorder against the trained human observer has been confirmed previously using the protocol of the British Hypertension Society and in this article the updated system incorporating a number of innovations is described. PMID:10234128

  13. Something for Everyone? An Evaluation of the Use of Audio-Visual Resources in Geographical Learning in the UK.

    ERIC Educational Resources Information Center

    McKendrick, John H.; Bowden, Annabel

    1999-01-01

    Reports from a survey of geographers that canvassed experiences using audio-visual resources to support teaching. Suggests that geographical learning has embraced audio-visual resources and that they are employed effectively. Concludes that integration of audio-visual resources into mainstream curriculum is essential to ensure effective and…

  14. Comparing the Effects of Classroom Audio-Recording and Video-Recording on Preservice Teachers' Reflection of Practice

    ERIC Educational Resources Information Center

    Bergman, Daniel

    2015-01-01

    This study examined the effects of audio and video self-recording on preservice teachers' written reflections. Participants (n = 201) came from a secondary teaching methods course and its school-based (clinical) fieldwork. The audio group (n[subscript A] = 106) used audio recorders to monitor their teaching in fieldwork placements; the video group…

  15. Audio-visual speech perception: a developmental ERP investigation

    PubMed Central

    Knowland, Victoria CP; Mercure, Evelyne; Karmiloff-Smith, Annette; Dick, Fred; Thomas, Michael SC

    2014-01-01

    Being able to see a talking face confers a considerable advantage for speech perception in adulthood. However, behavioural data currently suggest that children fail to make full use of these available visual speech cues until age 8 or 9. This is particularly surprising given the potential utility of multiple informational cues during language learning. We therefore explored this at the neural level. The event-related potential (ERP) technique has been used to assess the mechanisms of audio-visual speech perception in adults, with visual cues reliably modulating auditory ERP responses to speech. Previous work has shown congruence-dependent shortening of auditory N1/P2 latency and congruence-independent attenuation of amplitude in the presence of auditory and visual speech signals, compared to auditory alone. The aim of this study was to chart the development of these well-established modulatory effects over mid-to-late childhood. Experiment 1 employed an adult sample to validate a child-friendly stimulus set and paradigm by replicating previously observed effects of N1/P2 amplitude and latency modulation by visual speech cues; it also revealed greater attenuation of component amplitude given incongruent audio-visual stimuli, pointing to a new interpretation of the amplitude modulation effect. Experiment 2 used the same paradigm to map cross-sectional developmental change in these ERP responses between 6 and 11 years of age. The effect of amplitude modulation by visual cues emerged over development, while the effect of latency modulation was stable over the child sample. These data suggest that auditory ERP modulation by visual speech represents separable underlying cognitive processes, some of which show earlier maturation than others over the course of development. PMID:24176002

  16. Audio-visual speech perception: a developmental ERP investigation.

    PubMed

    Knowland, Victoria C P; Mercure, Evelyne; Karmiloff-Smith, Annette; Dick, Fred; Thomas, Michael S C

    2014-01-01

    Being able to see a talking face confers a considerable advantage for speech perception in adulthood. However, behavioural data currently suggest that children fail to make full use of these available visual speech cues until age 8 or 9. This is particularly surprising given the potential utility of multiple informational cues during language learning. We therefore explored this at the neural level. The event-related potential (ERP) technique has been used to assess the mechanisms of audio-visual speech perception in adults, with visual cues reliably modulating auditory ERP responses to speech. Previous work has shown congruence-dependent shortening of auditory N1/P2 latency and congruence-independent attenuation of amplitude in the presence of auditory and visual speech signals, compared to auditory alone. The aim of this study was to chart the development of these well-established modulatory effects over mid-to-late childhood. Experiment 1 employed an adult sample to validate a child-friendly stimulus set and paradigm by replicating previously observed effects of N1/P2 amplitude and latency modulation by visual speech cues; it also revealed greater attenuation of component amplitude given incongruent audio-visual stimuli, pointing to a new interpretation of the amplitude modulation effect. Experiment 2 used the same paradigm to map cross-sectional developmental change in these ERP responses between 6 and 11 years of age. The effect of amplitude modulation by visual cues emerged over development, while the effect of latency modulation was stable over the child sample. These data suggest that auditory ERP modulation by visual speech represents separable underlying cognitive processes, some of which show earlier maturation than others over the course of development. PMID:24176002

  17. Development of an audio-based virtual gaming environment to assist with navigation skills in the blind.

    PubMed

    Connors, Erin C; Yazzolino, Lindsay A; Sánchez, Jaime; Merabet, Lotfi B

    2013-01-01

    Audio-based Environment Simulator (AbES) is virtual environment software designed to improve real world navigation skills in the blind. Using only audio based cues and set within the context of a video game metaphor, users gather relevant spatial information regarding a building's layout. This allows the user to develop an accurate spatial cognitive map of a large-scale three-dimensional space that can be manipulated for the purposes of a real indoor navigation task. After game play, participants are then assessed on their ability to navigate within the target physical building represented in the game. Preliminary results suggest that early blind users were able to acquire relevant information regarding the spatial layout of a previously unfamiliar building as indexed by their performance on a series of navigation tasks. These tasks included path finding through the virtual and physical building, as well as a series of drop off tasks. We find that the immersive and highly interactive nature of the AbES software appears to greatly engage the blind user to actively explore the virtual environment. Applications of this approach may extend to larger populations of visually impaired individuals. PMID:23568182

  18. Audio representations of multi-channel EEG: a new tool for diagnosis of brain disorders

    PubMed Central

    Vialatte, François B; Dauwels, Justin; Musha, Toshimitsu; Cichocki, Andrzej

    2012-01-01

    Objective: The objective of this paper is to develop audio representations of electroencephalographic (EEG) multichannel signals, useful for medical practitioners and neuroscientists. The fundamental question explored in this paper is whether clinically valuable information contained in the EEG, not available from the conventional graphical EEG representation, might become apparent through audio representations. Methods and Materials: Music scores are generated from sparse time-frequency maps of EEG signals. Specifically, EEG signals of patients with mild cognitive impairment (MCI) and (healthy) control subjects are considered. Statistical differences in the audio representations of MCI patients and control subjects are assessed through mathematical complexity indexes as well as a perception test; in the latter, participants try to distinguish between audio sequences from MCI patients and control subjects. Results: Several characteristics of the audio sequences, including sample entropy, number of notes, and synchrony, are significantly different in MCI patients and control subjects (Mann-Whitney p < 0.01). Moreover, the participants of the perception test were able to accurately classify the audio sequences (89% correctly classified). Conclusions: The proposed audio representation of multi-channel EEG signals helps to understand the complex structure of EEG. Promising results were obtained on a clinical EEG data set. PMID:23383399

  19. Interactive MPEG-4 low-bit-rate speech/audio transmission over the Internet

    NASA Astrophysics Data System (ADS)

    Liu, Fang; Kim, JongWon; Kuo, C.-C. Jay

    1999-11-01

    The recently developed MPEG-4 technology enables the coding and transmission of natural and synthetic audio-visual data in the form of objects. In an effort to extend the object-based functionality of MPEG-4 to real-time Internet applications, architectural prototypes of multiplex layer and transport layer tailored for transmission of MPEG-4 data over IP are under debate among Internet Engineering Task Force (IETF), and MPEG-4 systems Ad Hoc group. In this paper, we present an architecture for interactive MPEG-4 speech/audio transmission system over the Internet. It utilities a framework of Real Time Streaming Protocol (RTSP) over Real-time Transport Protocol (RTP) to provide controlled, on-demand delivery of real time speech/audio data. Based on a client-server model, a couple of low bit-rate bit streams (real-time speech/audio, pre- encoded speech/audio) are multiplexed and transmitted via a single RTP channel to the receiver. The MPEG-4 Scene Description (SD) and Object Descriptor (OD) bit streams are securely sent through the RTSP control channel. Upon receiving, an initial MPEG-4 audio- visual scene is constructed after de-multiplexing, decoding of bit streams, and scene composition. A receiver is allowed to manipulate the initial audio-visual scene presentation locally, or interactively arrange scene changes by sending requests to the server. A server may also choose to update the client with new streams and list of contents for user selection.

  20. TECHNICAL NOTE: Portable audio electronics for impedance-based measurements in microfluidics

    NASA Astrophysics Data System (ADS)

    Wood, Paul; Sinton, David

    2010-08-01

    We demonstrate the use of audio electronics-based signals to perform on-chip electrochemical measurements. Cell phones and portable music players are examples of consumer electronics that are easily operated and are ubiquitous worldwide. Audio output (play) and input (record) signals are voltage based and contain frequency and amplitude information. A cell phone, laptop soundcard and two compact audio players are compared with respect to frequency response; the laptop soundcard provides the most uniform frequency response, while the cell phone performance is found to be insufficient. The audio signals in the common portable music players and laptop soundcard operate in the range of 20 Hz to 20 kHz and are found to be applicable, as voltage input and output signals, to impedance-based electrochemical measurements in microfluidic systems. Validated impedance-based measurements of concentration (0.1-50 mM), flow rate (2-120 µL min-1) and particle detection (32 µm diameter) are demonstrated. The prevailing, lossless, wave audio file format is found to be suitable for data transmission to and from external sources, such as a centralized lab, and the cost of all hardware (in addition to audio devices) is ~10 USD. The utility demonstrated here, in combination with the ubiquitous nature of portable audio electronics, presents new opportunities for impedance-based measurements in portable microfluidic systems.

  1. Laboratory and in-flight experiments to evaluate 3-D audio display technology

    NASA Astrophysics Data System (ADS)

    Ericson, Mark; McKinley, Richard; Kibbe, Marion; Francis, Daniel

    1994-01-01

    Laboratory and in-flight experiments were conducted to evaluate 3-D audio display technology for cockpit applications. A 3-D audio display generator was developed which digitally encodes naturally occurring direction information onto any audio signal and presents the binaural sound over headphones. The acoustic image is stabilized for head movement by use of an electromagnetic head-tracking device. In the laboratory, a 3-D audio display generator was used to spatially separate competing speech messages to improve the intelligibility of each message. Up to a 25 percent improvement in intelligibility was measured for spatially separated speech at high ambient noise levels (115 dB SPL). During the in-flight experiments, pilots reported that spatial separation of speech communications provided a noticeable improvement in intelligibility. The use of 3-D audio for target acquisition was also investigated. In the laboratory, 3-D audio enabled the acquisition of visual targets in about two seconds average response time at 17 degrees accuracy. During the in-flight experiments, pilots correctly identified ground targets 50, 75, and 100 percent of the time at separation angles of 12, 20, and 35 degrees, respectively. In general, pilot performance in the field with the 3-D audio display generator was as expected, based on data from laboratory experiments.

  2. Incorporating Data Link Features into a Multi-Function Display to Support Self-Separation and Spacing Tasks for General Aviation Pilots

    NASA Technical Reports Server (NTRS)

    Adams, Catherine A.; Murdoch, Jennifer L.; Consiglio, Maria C.; WIlliams, Daniel M.

    2005-01-01

    One objective of the Small Aircraft Transportation System (SATS) Higher Volume Operations (HVO) project is to increase the capacity and utilization of small non-towered, non-radar equipped airports by transferring traffic management activities to an automated Airport Management Module (AMM) and separation responsibilities to general aviation (GA) pilots. Implementation of this concept required the development of a research Multi-Function Display (MFD) to support the interactive communications between pilots and the AMM. The interface also had to accommodate traffic awareness, self-separation, and spacing tasks through dynamic messaging and symbology for flight path conformance and conflict detection and alerting (CDA). The display served as the mechanism to support the examination of the viability of executing instrument operations designed for SATS designated airports. Results of simulation and flight experiments conducted at the National Aeronautics and Space Administration's (NASA) Langley Research Center indicate that the concept, as facilitated by the research MFD, did not increase pilots subjective workload levels or reduce their situation awareness (SA). Post-test usability assessments revealed that pilots preferred using the enhanced MFD to execute flight procedures, reporting improved SA over conventional instrument flight rules (IFR) procedures.

  3. Method for Reading Sensors and Controlling Actuators Using Audio Interfaces of Mobile Devices

    PubMed Central

    Aroca, Rafael V.; Burlamaqui, Aquiles F.; Gonçalves, Luiz M. G.

    2012-01-01

    This article presents a novel closed loop control architecture based on audio channels of several types of computing devices, such as mobile phones and tablet computers, but not restricted to them. The communication is based on an audio interface that relies on the exchange of audio tones, allowing sensors to be read and actuators to be controlled. As an application example, the presented technique is used to build a low cost mobile robot, but the system can also be used in a variety of mechatronics applications and sensor networks, where smartphones are the basic building blocks. PMID:22438726

  4. When the third party observer of a neuropsychological evaluation is an audio-recorder.

    PubMed

    Constantinou, Marios; Ashendorf, Lee; McCaffrey, Robert J

    2002-08-01

    The presence of third parties during neuropsychological evaluations is an issue of concern for contemporary neuropsychologists. Previous studies have reported that the presence of an observer during neuropsychological testing alters the performance of individuals under evaluation. The present study sought to investigate whether audio-recording affects the neuropsychological test performance of individuals in the same way that third party observation does. In the presence of an audio-recorder the performance of the participants on memory tests declined. Performance on motor tests, on the other hand, was not affected by the presence of an audio-recorder. The implications of these findings in forensic neuropsychological evaluations are discussed. PMID:12607152

  5. The cortical representation of the speech envelope is earlier for audiovisual speech than audio speech.

    PubMed

    Crosse, Michael J; Lalor, Edmund C

    2014-04-01

    Visual speech can greatly enhance a listener's comprehension of auditory speech when they are presented simultaneously. Efforts to determine the neural underpinnings of this phenomenon have been hampered by the limited temporal resolution of hemodynamic imaging and the fact that EEG and magnetoencephalographic data are usually analyzed in response to simple, discrete stimuli. Recent research has shown that neuronal activity in human auditory cortex tracks the envelope of natural speech. Here, we exploit this finding by estimating a linear forward-mapping between the speech envelope and EEG data and show that the latency at which the envelope of natural speech is represented in cortex is shortened by >10 ms when continuous audiovisual speech is presented compared with audio-only speech. In addition, we use a reverse-mapping approach to reconstruct an estimate of the speech stimulus from the EEG data and, by comparing the bimodal estimate with the sum of the unimodal estimates, find no evidence of any nonlinear additive effects in the audiovisual speech condition. These findings point to an underlying mechanism that could account for enhanced comprehension during audiovisual speech. Specifically, we hypothesize that low-level acoustic features that are temporally coherent with the preceding visual stream may be synthesized into a speech object at an earlier latency, which may provide an extended period of low-level processing before extraction of semantic information. PMID:24401714

  6. Audio Classification in Speech and Music: A Comparison between a Statistical and a Neural Approach

    NASA Astrophysics Data System (ADS)

    Bugatti, Alessandro; Flammini, Alessandra; Migliorati, Pierangelo

    2002-12-01

    We focus the attention on the problem of audio classification in speech and music for multimedia applications. In particular, we present a comparison between two different techniques for speech/music discrimination. The first method is based on Zero crossing rate and Bayesian classification. It is very simple from a computational point of view, and gives good results in case of pure music or speech. The simulation results show that some performance degradation arises when the music segment contains also some speech superimposed on music, or strong rhythmic components. To overcome these problems, we propose a second method, that uses more features, and is based on neural networks (specifically a multi-layer Perceptron). In this case we obtain better performance, at the expense of a limited growth in the computational complexity. In practice, the proposed neural network is simple to be implemented if a suitable polynomial is used as the activation function, and a real-time implementation is possible even if low-cost embedded systems are used.

  7. Do audio-visual motion cues promote segregation of auditory streams?

    PubMed

    Shestopalova, Lidia; Bőhm, Tamás M; Bendixen, Alexandra; Andreou, Andreas G; Georgiou, Julius; Garreau, Guillaume; Hajdu, Botond; Denham, Susan L; Winkler, István

    2014-01-01

    An audio-visual experiment using moving sound sources was designed to investigate whether the analysis of auditory scenes is modulated by synchronous presentation of visual information. Listeners were presented with an alternating sequence of two pure tones delivered by two separate sound sources. In different conditions, the two sound sources were either stationary or moving on random trajectories around the listener. Both the sounds and the movement trajectories were derived from recordings in which two humans were moving with loudspeakers attached to their heads. Visualized movement trajectories modeled by a computer animation were presented together with the sounds. In the main experiment, behavioral reports on sound organization were collected from young healthy volunteers. The proportion and stability of the different sound organizations were compared between the conditions in which the visualized trajectories matched the movement of the sound sources and when the two were independent of each other. The results corroborate earlier findings that separation of sound sources in space promotes segregation. However, no additional effect of auditory movement per se on the perceptual organization of sounds was obtained. Surprisingly, the presentation of movement-congruent visual cues did not strengthen the effects of spatial separation on segregating auditory streams. Our findings are consistent with the view that bistability in the auditory modality can occur independently from other modalities. PMID:24778604

  8. Emerging magnetic technologies for consumer audio/video (invited)

    NASA Astrophysics Data System (ADS)

    Fujiwara, Hideo

    1993-05-01

    In the field of consumer audio/video, digital technology is the natural path for advancement. In audio systems, it has just been introduced in the form of digital compact cassette tape recorder and mini disk system in which magneto-optical recording is used. Therefore, the digital video instruments, such as small cassette digital video tape recorders (VTR) for high-definition television and static or short moving image devices will be the next attractive goals to be achieved. Technology analysis shows that the basic techniques of signal processing for small cassette VTR, in which the recording density of 1 μm2/bit and the data rate of 130-160 Mb/s are required, are almost at hand. For the media, improved metal-evaporated tape including oblique incident Co-CoO will be the most favorable candidate, although particulate media of metal powder tape (including Co or N modified forms) or Ba-ferrite tape coated by using the newly developed ultrathin coating techniques may take its position with the help of enhanced head sensitivity, because of their better durability. The basic demand for increasing saturation induction Bs of the head materials will be fulfilled by fine grain Fe-based materials, the available Bs being as high as 2 T. A dramatic increase in magnetoresistive (MR) head sensitivity is expected by using the giant MR effect, although a breakthrough is required to extract its full potential. On the other hand, varieties of static (including short moving) image devices will be developed by use of rapidly progressing magneto-optical recording technology. Preliminary techniques are now ready to meet the demand for the required increase in bit rate and recording density. A bit rate of 64 Mbits/s (8 Mbytes/s) has already been attained. A bit density of 107 bits/mm2 (6.5 Gb/in.2) is likely to be realized, and a factor of 2 or 3 more enhancement can be counted on. Rigid disk and perpendicular recording will have a chance to participate in the field if their techniques

  9. A high-resolution atlas of the infrared spectrum of the Sun and the Earth atmosphere from space. Volume 3: Key to identification of solar features

    NASA Technical Reports Server (NTRS)

    Geller, Murray

    1992-01-01

    During the period April 29 through May 2, 1985, the Atmospheric Trace Molecule Spectroscopy (ATMOS) experiment was operated as part of the Spacelab-3 (SL-3) payload on the shuttle Challenger. The instrument, a Fourier transform spectrometer, recorded over 2000 infrared solar spectra from an altitude of 360 km. Although the majority of the spectra were taken through the limb of the Earth's atmosphere in order to better understand its composition, several hundred of the 'high-sun' spectra were completely free from telluric absorption. These high-sun spectra recorded from space are, at the present time, the only high-resolution infrared spectra ever taken of the Sun free from absorptions due to constituents in the Earth's atmosphere. Volumes 1 and 2 of this series provide a compilation of these spectra arranged in a format suitable for quick-look reference purposes and are the first record of the continuous high-resolution infrared spectrum of the Sun and the Earth's atmosphere from space. In the Table of Identifications, which constitutes the main body of this volume, each block of eight wavenumbers is given a separate heading and corresponds to a page of two panels in Volume 1 of this series. In addition, three separate blocks of data available from ATMOS from 622-630 cm(exp -1), 630-638 cm(exp -1) and 638-646 cm(exp -1), excluded from Volume 1 because of the low signal-to-noise ratio, have been included due to the certain identification of several OH and NH transitions. In the first column of the table, the corrected frequency is given. The second column identifies the molecular species. The third and fourth columns represent the assigned transition. The fifth column gives the depth of the molecular line in millimeters. Also included in this column is a notation to indicate whether the line is a blend or lies on the shoulder(s) of another line(s). The final column repeats a question mark if the line is unidentified.

  10. Testing of a Methane Cryogenic Heat Pipe with a Liquid Trap Turn-Off Feature for use on Space Interferometer Mission (SIM)

    NASA Technical Reports Server (NTRS)

    Cepeda-Rizo, Juan; Krylo, Robert; Fisher, Melanie; Bugby, David C.

    2011-01-01

    Camera cooling for SIM presents three thermal control challenges; stable operation at 163K (110 C), decontamination heating to +20 C, and a long span from the cameras to the radiator. A novel cryogenic cooling system based on a methane heat pipe meets these challenges. The SIM thermal team, with the help of heat pipe vendor ATK, designed and tested a complete, low temperature, cooling system. The system accommodates the two SIM cameras with a double-ended conduction bar, a single methane heat pipe, independent turn-off devices, and a flight-like radiator. The turn ]off devices consist of a liquid trap, for removing the methane from the pipe, and an electrical heater to raise the methane temperature above the critical point thus preventing two-phase operation. This is the first time a cryogenic heat pipe has been tested at JPL and is also the first heat pipe to incorporate the turn-off features. Operation at 163K with a methane heat pipe is an important new thermal control capability for the lab. In addition, the two turn-off technologies enhance the "bag of tricks" available to the JPL thermal community. The successful test program brings this heat pipe to a high level of technology readiness.

  11. Effects of audio-visual stimulation on the incidence of restraint ulcers on the Wistar rat

    NASA Technical Reports Server (NTRS)

    Martin, M. S.; Martin, F.; Lambert, R.

    1979-01-01

    The role of sensory simulation in restrained rats was investigated. Both mixed audio-visual and pure sound stimuli, ineffective in themselves, were found to cause a significant increase in the incidence of restraint ulcers in the Wistar Rat.

  12. Worldwide survey of direct-to-listener digital audio delivery systems development since WARC-1992

    NASA Technical Reports Server (NTRS)

    Messer, Dion D.

    1993-01-01

    Each country was allocated frequency band(s) for direct-to-listener digital audio broadcasting at WARC-92. These allocations were near 1500, 2300, and 2600 MHz. In addition, some countries are encouraging the development of digital audio broadcasting services for terrestrial delivery only in the VHF bands (at frequencies from roughly 50 to 300 MHz) and in the medium-wave broadcasting band (AM band) (from roughly 0.5 to 1.7 MHz). The development activity increase was explosive. Current development, as of February 1993, as it is known to the author is summarized. The information given includes the following characteristics, as appropriate, for each planned system: coverage areas, audio quality, number of audio channels, delivery via satellite/terrestrial or both, carrier frequency bands, modulation methods, source coding, and channel coding. Most proponents claim that they will be operational in 3 or 4 years.

  13. 37 CFR 202.22 - Acquisition and deposit of unpublished audio and audiovisual transmission programs.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... Copyrights COPYRIGHT OFFICE, LIBRARY OF CONGRESS COPYRIGHT OFFICE AND PROCEDURES PREREGISTRATION AND... and copies of unpublished audio and audiovisual transmission programs by the Library of Congress under... transmission programs. (1) Library of Congress employees, including Library of Congress contractors,...

  14. Effectiveness and Comparison of Various Audio Distraction Aids in Management of Anxious Dental Paediatric Patients

    PubMed Central

    Johri, Nikita; Khan, Suleman Abbas; Singh, Rahul Kumar; Chadha, Dheera; Navit, Pragati; Sharma, Anshul; Bahuguna, Rachana

    2015-01-01

    Background Dental anxiety is a widespread phenomenon and a concern for paediatric dentistry. The inability of children to deal with threatening dental stimuli often manifests as behaviour management problems. Nowadays, the use of non-aversive behaviour management techniques is more advocated, which are more acceptable to parents, patients and practitioners. Therefore, this present study was conducted to find out which audio aid was the most effective in the managing anxious children. Aims and Objectives The aim of the present study was to compare the efficacy of audio-distraction aids in reducing the anxiety of paediatric patients while undergoing various stressful and invasive dental procedures. The objectives were to ascertain whether audio distraction is an effective means of anxiety management and which type of audio aid is the most effective. Materials and Methods A total number of 150 children, aged between 6 to 12 years, randomly selected amongst the patients who came for their first dental check-up, were placed in five groups of 30 each. These groups were the control group, the instrumental music group, the musical nursery rhymes group, the movie songs group and the audio stories group. The control group was treated under normal set-up & audio group listened to various audio presentations during treatment. Each child had four visits. In each visit, after the procedures was completed, the anxiety levels of the children were measured by the Venham’s Picture Test (VPT), Venham’s Clinical Rating Scale (VCRS) and pulse rate measurement with the help of pulse oximeter. Results A significant difference was seen between all the groups for the mean pulse rate, with an increase in subsequent visit. However, no significant difference was seen in the VPT & VCRS scores between all the groups. Audio aids in general reduced anxiety in comparison to the control group, and the most significant reduction in anxiety level was observed in the audio stories group

  15. Integration of advanced teleoperation technologies for control of space robots

    NASA Technical Reports Server (NTRS)

    Stagnaro, Michael J.

    1993-01-01

    Teleoperated robots require one or more humans to control actuators, mechanisms, and other robot equipment given feedback from onboard sensors. To accomplish this task, the human or humans require some form of control station. Desirable features of such a control station include operation by a single human, comfort, and natural human interfaces (visual, audio, motion, tactile, etc.). These interfaces should work to maximize performance of the human/robot system by streamlining the link between human brain and robot equipment. This paper describes development of a control station testbed with the characteristics described above. Initially, this testbed will be used to control two teleoperated robots. Features of the robots include anthropomorphic mechanisms, slaving to the testbed, and delivery of sensory feedback to the testbed. The testbed will make use of technologies such as helmet mounted displays, voice recognition, and exoskeleton masters. It will allow tor integration and testing of emerging telepresence technologies along with techniques for coping with control link time delays. Systems developed from this testbed could be applied to ground control of space based robots. During man-tended operations, the Space Station Freedom may benefit from ground control of IVA or EVA robots with science or maintenance tasks. Planetary exploration may also find advanced teleoperation systems to be very useful.

  16. Audio watermarking forensics: detecting malicious re-embedding

    NASA Astrophysics Data System (ADS)

    Zmudzinski, Sascha; Steinebach, Martin; Katzenbeisser, Stefan; Rührmair, Ulrich

    2010-01-01

    Digital watermarking has become a widely used security technology in the domain of digital rights management and copyright protection as well as in other applications. In this work, we show recent results regarding a particular security attack: Embedding a new message in a previously watermarked cover using the same key as the original message. This re-embedding can be the consequence of the absence of truly asymmetric watermarking solutions, especially if the watermark is to be detected in public. In public detection scenarios, every detector needs the same key the embedder used to watermark the cover. With knowledge of the embedding algorithm, everybody who is able to detect the message can also maliciously embed a new message with the same key over the old one. This scenario is relevant in the case that an attacker intends to counterfeit a copyright notice, transaction ID or to change an embedded authentication code. This work presents experimental results on mechanisms for identifying such multiple embeddings in a spreadspectrum patchwork audio watermarking approach. We demonstrate that under certain circumstances such multiple embedding can be detected by watermarking-forensics.

  17. Audio-visual aid in teaching "fatty liver".

    PubMed

    Dash, Sambit; Kamath, Ullas; Rao, Guruprasad; Prakash, Jay; Mishra, Snigdha

    2016-05-01

    Use of audio visual tools to aid in medical education is ever on a rise. Our study intends to find the efficacy of a video prepared on "fatty liver," a topic that is often a challenge for pre-clinical teachers, in enhancing cognitive processing and ultimately learning. We prepared a video presentation of 11:36 min, incorporating various concepts of the topic, while keeping in view Mayer's and Ellaway guidelines for multimedia presentation. A pre-post test study on subject knowledge was conducted for 100 students with the video shown as intervention. A retrospective pre study was conducted as a survey which inquired about students understanding of the key concepts of the topic and a feedback on our video was taken. Students performed significantly better in the post test (mean score 8.52 vs. 5.45 in pre-test), positively responded in the retrospective pre-test and gave a positive feedback for our video presentation. Well-designed multimedia tools can aid in cognitive processing and enhance working memory capacity as shown in our study. In times when "smart" device penetration is high, information and communication tools in medical education, which can act as essential aid and not as replacement for traditional curriculums, can be beneficial to the students. © 2015 by The International Union of Biochemistry and Molecular Biology, 44:241-245, 2016. PMID:26625860

  18. Frequency allocations for a new satellite service - Digital audio broadcasting

    NASA Technical Reports Server (NTRS)

    Reinhart, Edward E.

    1992-01-01

    The allocation in the range 500-3000 MHz for digital audio broadcasting (DAB) is described in terms of key issues such as the transmission-system architectures. Attention is given to the optimal amount of spectrum for allocation and the technological considerations relevant to downlink bands for satellite and terrestrial transmissions. Proposals for DAB allocations are compared, and reference is made to factors impinging on the provision of ground/satellite feeder links. The allocation proposals describe the implementation of 50-60-MHz bandwidths for broadcasting in the ranges near 800 MHz, below 1525 MHz, near 2350 MHz, and near 2600 MHz. Three specific proposals are examined in terms of characteristics such as service areas, coverage/beam, channels/satellite beam, and FCC license status. Several existing problems are identified including existing services crowded with systems, the need for new bands in the 1000-3000-MHz range, and variations in the nature and intensity of implementations of existing allocations that vary from country to country.

  19. Interactive video audio system: communication server for INDECT portal

    NASA Astrophysics Data System (ADS)

    Mikulec, Martin; Voznak, Miroslav; Safarik, Jakub; Partila, Pavol; Rozhon, Jan; Mehic, Miralem

    2014-05-01

    The paper deals with presentation of the IVAS system within the 7FP EU INDECT project. The INDECT project aims at developing the tools for enhancing the security of citizens and protecting the confidentiality of recorded and stored information. It is a part of the Seventh Framework Programme of European Union. We participate in INDECT portal and the Interactive Video Audio System (IVAS). This IVAS system provides a communication gateway between police officers working in dispatching centre and police officers in terrain. The officers in dispatching centre have capabilities to obtain information about all online police officers in terrain, they can command officers in terrain via text messages, voice or video calls and they are able to manage multimedia files from CCTV cameras or other sources, which can be interesting for officers in terrain. The police officers in terrain are equipped by smartphones or tablets. Besides common communication, they can reach pictures or videos sent by commander in office and they can respond to the command via text or multimedia messages taken by their devices. Our IVAS system is unique because we are developing it according to the special requirements from the Police of the Czech Republic. The IVAS communication system is designed to use modern Voice over Internet Protocol (VoIP) services. The whole solution is based on open source software including linux and android operating systems. The technical details of our solution are presented in the paper.

  20. Audio-visual assistance in co-creating transition knowledge

    NASA Astrophysics Data System (ADS)

    Hezel, Bernd; Broschkowski, Ephraim; Kropp, Jürgen P.

    2013-04-01

    Earth system and climate impact research results point to the tremendous ecologic, economic and societal implications of climate change. Specifically people will have to adopt lifestyles that are very different from those they currently strive for in order to mitigate severe changes of our known environment. It will most likely not suffice to transfer the scientific findings into international agreements and appropriate legislation. A transition is rather reliant on pioneers that define new role models, on change agents that mainstream the concept of sufficiency and on narratives that make different futures appealing. In order for the research community to be able to provide sustainable transition pathways that are viable, an integration of the physical constraints and the societal dynamics is needed. Hence the necessary transition knowledge is to be co-created by social and natural science and society. To this end, the Climate Media Factory - in itself a massively transdisciplinary venture - strives to provide an audio-visual connection between the different scientific cultures and a bi-directional link to stake holders and society. Since methodology, particular language and knowledge level of the involved is not the same, we develop new entertaining formats on the basis of a "complexity on demand" approach. They present scientific information in an integrated and entertaining way with different levels of detail that provide entry points to users with different requirements. Two examples shall illustrate the advantages and restrictions of the approach.

  1. Differentiated audio-tactile correspondences in sighted and blind individuals.

    PubMed

    Deroy, Ophelia; Fasiello, Irène; Hayward, Vincent; Auvray, Malika

    2016-08-01

    The aim of the present study is to investigate whether the crossmodal correspondence robustly documented between auditory pitch and visual elevation has analogues in the audio-tactile domain. Across 4 experiments, the compatibility effects between intuitively congruent pairs of stimuli (i.e., outward tactile movement, going from the inside of the finger toward the fingertip and increasing pitch, or inward tactile movement and decreasing pitch) and incongruent pairs stimuli (i.e., the reverse associations) were measured. Two methods were compared to assess the behavioral effects of such a correspondence: One where participants have to respond to either the auditory or tactile stimulus presented simultaneously, while ignoring the other (speeded classification task), and the other where the auditory and tactile stimuli are presented sequentially and associated to different response buttons (implicit association test). No significant compatibility effect was observed under the speeded classification task. The implicit association test revealed a significant compatibility effect. This effect was similar in the conditions where the finger was placed vertically and horizontally. However, this implicit association between pitch and tactile movements was not observed in blind participants. These results have methodological implications for the explanation and testing of crossmodal correspondences, and the origin of the widely discussed association between pitch and vertical elevation. (PsycINFO Database Record PMID:26950385

  2. Comparing Learning Gains: Audio Versus Text-based Instructor Communication in a Blended Online Learning Environment

    NASA Astrophysics Data System (ADS)

    Shimizu, Dominique

    Though blended course audio feedback has been associated with several measures of course satisfaction at the postsecondary and graduate levels compared to text feedback, it may take longer to prepare and positive results are largely unverified in K-12 literature. The purpose of this quantitative study was to investigate the time investment and learning impact of audio communications with 228 secondary students in a blended online learning biology unit at a central Florida public high school. A short, individualized audio message regarding the student's progress was given to each student in the audio group; similar text-based messages were given to each student in the text-based group on the same schedule; a control got no feedback. A pretest and posttest were employed to measure learning gains in the three groups. To compare the learning gains in two types of feedback with each other and to no feedback, a controlled, randomized, experimental design was implemented. In addition, the creation and posting of audio and text feedback communications were timed in order to assess whether audio feedback took longer to produce than text only feedback. While audio feedback communications did take longer to create and post, there was no difference between learning gains as measured by posttest scores when student received audio, text-based, or no feedback. Future studies using a similar randomized, controlled experimental design are recommended to verify these results and test whether the trend holds in a broader range of subjects, over different time frames, and using a variety of assessment types to measure student learning.

  3. Using the ENF Criterion for Determining the Time of Recording of Short Digital Audio Recordings

    NASA Astrophysics Data System (ADS)

    Huijbregtse, Maarten; Geradts, Zeno

    The Electric Network Frequency (ENF) Criterion is a recently developed forensic technique for determining the time of recording of digital audio recordings, by matching the ENF pattern from a questioned recording with an ENF pattern database. In this paper we discuss its inherent limitations in the case of short - i.e., less than 10 minutes in duration - digital audio recordings. We also present a matching procedure based on the correlation coefficient, as a more robust alternative to squared error matching.

  4. Sounds of Space

    NASA Astrophysics Data System (ADS)

    Gurnett, D. A.

    2005-12-01

    Starting in the early 1960s, spacecraft-borne plasma wave instruments revealed that space is filled with an astonishing variety of radio and plasma wave sounds, which have come to be called "sounds of space." For over forty years these sounds have been collected and played to a wide variety of audiences, often as the result of press conferences or press releases involving various NASA projects for which the University of Iowa has provided plasma wave instruments. This activity has led to many interviews on local and national radio programs, and occasionally on programs haviang world-wide coverage, such as the BBC. As a result of this media coverage, we have been approached many times by composers requesting copies of our space sounds for use in their various projects, many of which involve electronic synthesis of music. One of these collaborations led to "Sun Rings," which is a musical event produced by the Kronos Quartet that has played to large audiences all over the world. With the availability of modern computer graphic techniques we have recently been attempting to integrate some of these sound of space into an educational audio/video web site that illustrates the scientific principles involved in the origin of space plasma waves. Typically I try to emphasize that a substantial gas pressure exists everywhere in space in the form of an ionized gas called a plasma, and that this plasma can lead to a wide variety of wave phenomenon. Examples of some of this audio/video material will be presented.

  5. StirMark Benchmark: audio watermarking attacks based on lossy compression

    NASA Astrophysics Data System (ADS)

    Steinebach, Martin; Lang, Andreas; Dittmann, Jana

    2002-04-01

    StirMark Benchmark is a well-known evaluation tool for watermarking robustness. Additional attacks are added to it continuously. To enable application based evaluation, in our paper we address attacks against audio watermarks based on lossy audio compression algorithms to be included in the test environment. We discuss the effect of different lossy compression algorithms like MPEG-2 audio Layer 3, Ogg or VQF on a selection of audio test data. Our focus is on changes regarding the basic characteristics of the audio data like spectrum or average power and on removal of embedded watermarks. Furthermore we compare results of different watermarking algorithms and show that lossy compression is still a challenge for most of them. There are two strategies for adding evaluation of robustness against lossy compression to StirMark Benchmark: (a) use of existing free compression algorithms (b) implementation of a generic lossy compression simulation. We discuss how such a model can be implemented based on the results of our tests. This method is less complex, as no real psycho acoustic model has to be applied. Our model can be used for audio watermarking evaluation of numerous application fields. As an example, we describe its importance for e-commerce applications with watermarking security.

  6. Audio-vocal interaction in single neurons of the monkey ventrolateral prefrontal cortex.

    PubMed

    Hage, Steffen R; Nieder, Andreas

    2015-05-01

    Complex audio-vocal integration systems depend on a strong interconnection between the auditory and the vocal motor system. To gain cognitive control over audio-vocal interaction during vocal motor control, the PFC needs to be involved. Neurons in the ventrolateral PFC (VLPFC) have been shown to separately encode the sensory perceptions and motor production of vocalizations. It is unknown, however, whether single neurons in the PFC reflect audio-vocal interactions. We therefore recorded single-unit activity in the VLPFC of rhesus monkeys (Macaca mulatta) while they produced vocalizations on command or passively listened to monkey calls. We found that 12% of randomly selected neurons in VLPFC modulated their discharge rate in response to acoustic stimulation with species-specific calls. Almost three-fourths of these auditory neurons showed an additional modulation of their discharge rates either before and/or during the monkeys' motor production of vocalization. Based on these audio-vocal interactions, the VLPFC might be well positioned to combine higher order auditory processing with cognitive control of the vocal motor output. Such audio-vocal integration processes in the VLPFC might constitute a precursor for the evolution of complex learned audio-vocal integration systems, ultimately giving rise to human speech. PMID:25948255

  7. An Audio Architecture Integrating Sound and Live Voice for Virtual Environments

    NASA Astrophysics Data System (ADS)

    Krebs, Eric M.

    2002-09-01

    The purpose behind this thesis was to design and implement audio system architecture, both in hardware and in software, for use in virtual environments The hardware and software design requirements were aimed at implementing acoustical models, such as reverberation and occlusion, and live audio streaming to any simulation employing this architecture, Several free or open-source sound APIs were evaluated, and DirectSound3DTM was selected as the core component of the audio architecture, Creative Technology Ltd, Environmental Audio Extensions (EAXTM 3,0) were integrated into the architecture to provide environmental effects such as reverberation, occlusion, obstruction, and exclusion, Voice over IP (VoIP) technology was evaluated to provide live, streaming voice to any virtual environment DirectVoice was selected as the voice component of the VoIP architecture due to its integration with DirectSound3DTM, However, extremely high latency considerations with DirectVoice, and any other VoIP application or software, required further research into alternative live voice architectures for inclusion in virtual environments Ausim3D's GoldServe Audio System was evaluated and integrated into the hardware component of the audio architecture to provide an extremely low-latency, live, streaming voice capability.

  8. Do gender differences in audio-visual benefit and visual influence in audio-visual speech perception emerge with age?

    PubMed Central

    Alm, Magnus; Behne, Dawn

    2015-01-01

    Gender and age have been found to affect adults’ audio-visual (AV) speech perception. However, research on adult aging focuses on adults over 60 years, who have an increasing likelihood for cognitive and sensory decline, which may confound positive effects of age-related AV-experience and its interaction with gender. Observed age and gender differences in AV speech perception may also depend on measurement sensitivity and AV task difficulty. Consequently both AV benefit and visual influence were used to measure visual contribution for gender-balanced groups of young (20–30 years) and middle-aged adults (50–60 years) with task difficulty varied using AV syllables from different talkers in alternative auditory backgrounds. Females had better speech-reading performance than males. Whereas no gender differences in AV benefit or visual influence were observed for young adults, visually influenced responses were significantly greater for middle-aged females than middle-aged males. That speech-reading performance did not influence AV benefit may be explained by visual speech extraction and AV integration constituting independent abilities. Contrastingly, the gender difference in visually influenced responses in middle adulthood may reflect an experience-related shift in females’ general AV perceptual strategy. Although young females’ speech-reading proficiency may not readily contribute to greater visual influence, between young and middle-adulthood recurrent confirmation of the contribution of visual cues induced by speech-reading proficiency may gradually shift females AV perceptual strategy toward more visually dominated responses. PMID:26236274

  9. New Design Feature in Virginia's Public Schools.

    ERIC Educational Resources Information Center

    Gillespie, Monica McTeague

    A study of new design features for public schools was conducted to determine what specific design features were being implemented in Virginia's schools. This report summarizes the major trends in facility design that were discovered in the following areas: safety and security features; classroom space; technology areas; administrative spaces;…

  10. A prototype feature system for feature retrieval using relationships

    USGS Publications Warehouse

    Choi, J.; Usery, E.L.

    2009-01-01

    Using a feature data model, geographic phenomena can be represented effectively by integrating space, theme, and time. This paper extends and implements a feature data model that supports query and visualization of geographic features using their non-spatial and temporal relationships. A prototype feature-oriented geographic information system (FOGIS) is then developed and storage of features named Feature Database is designed. Buildings from the U.S. Marine Corps Base, Camp Lejeune, North Carolina and subways in Chicago, Illinois are used to test the developed system. The results of the applications show the strength of the feature data model and the developed system 'FOGIS' when they utilize non-spatial and temporal relationships in order to retrieve and visualize individual features.

  11. The Black Record: A Selective Discography of Afro-Americana on Audio Discs Held by the Audio/Visual Department, John M. Olin Library.

    ERIC Educational Resources Information Center

    Dain, Bernice, Comp.; Nevin, David, Comp.

    The present revised and expanded edition of this document is an inclusive cumulation. A few items have been included which are on order as new to the collection or as replacements. This discography is intended to serve primarily as a local user's guide. The call number preceding each entry is based on the Audio-Visual Department's own, unique…

  12. Dialect Features in the Language of Black Characters on American Television Programming.

    ERIC Educational Resources Information Center

    Fine, Marlene G.; Anderson, Carolyn

    This study describes the syntactic features of Black English Vernacular (BEV) spoken by black characters in three black situation comedies on American television: "The Jeffersons,""Good Times," and "What's Happening." Using scripts and audio tapes of three episodes from each series during the 1977-78 television season, transcripts were made of…

  13. The Development of Audio-Visual Integration for Temporal Judgements

    PubMed Central

    Adams, Wendy J.

    2016-01-01

    Adults combine information from different sensory modalities to estimate object properties such as size or location. This process is optimal in that (i) sensory information is weighted according to relative reliability: more reliable estimates have more influence on the combined estimate and (ii) the combined estimate is more reliable than the component uni-modal estimates. Previous studies suggest that optimal sensory integration does not emerge until around 10 years of age. Younger children rely on a single modality or combine information using inappropriate sensory weights. Children aged 4–11 and adults completed a simple audio-visual task in which they reported either the number of beeps or the number of flashes in uni-modal and bi-modal conditions. In bi-modal trials, beeps and flashes differed in number by 0, 1 or 2. Mutual interactions between the sensory signals were evident at all ages: the reported number of flashes was influenced by the number of simultaneously presented beeps and vice versa. Furthermore, for all ages, the relative strength of these interactions was predicted by the relative reliabilities of the two modalities, in other words, all observers weighted the signals appropriately. The degree of cross-modal interaction decreased with age: the youngest observers could not ignore the task-irrelevant modality—they fully combined vision and audition such that they perceived equal numbers of flashes and beeps for bi-modal stimuli. Older observers showed much smaller effects of the task-irrelevant modality. Do these interactions reflect optimal integration? Full or partial cross-modal integration predicts improved reliability in bi-modal conditions. In contrast, switching between modalities reduces reliability. Model comparison suggests that older observers employed partial integration, whereas younger observers (up to around 8 years) did not integrate, but followed a sub-optimal switching strategy, responding according to either visual or

  14. The Development of Audio-Visual Integration for Temporal Judgements.

    PubMed

    Adams, Wendy J

    2016-04-01

    Adults combine information from different sensory modalities to estimate object properties such as size or location. This process is optimal in that (i) sensory information is weighted according to relative reliability: more reliable estimates have more influence on the combined estimate and (ii) the combined estimate is more reliable than the component uni-modal estimates. Previous studies suggest that optimal sensory integration does not emerge until around 10 years of age. Younger children rely on a single modality or combine information using inappropriate sensory weights. Children aged 4-11 and adults completed a simple audio-visual task in which they reported either the number of beeps or the number of flashes in uni-modal and bi-modal conditions. In bi-modal trials, beeps and flashes differed in number by 0, 1 or 2. Mutual interactions between the sensory signals were evident at all ages: the reported number of flashes was influenced by the number of simultaneously presented beeps and vice versa. Furthermore, for all ages, the relative strength of these interactions was predicted by the relative reliabilities of the two modalities, in other words, all observers weighted the signals appropriately. The degree of cross-modal interaction decreased with age: the youngest observers could not ignore the task-irrelevant modality-they fully combined vision and audition such that they perceived equal numbers of flashes and beeps for bi-modal stimuli. Older observers showed much smaller effects of the task-irrelevant modality. Do these interactions reflect optimal integration? Full or partial cross-modal integration predicts improved reliability in bi-modal conditions. In contrast, switching between modalities reduces reliability. Model comparison suggests that older observers employed partial integration, whereas younger observers (up to around 8 years) did not integrate, but followed a sub-optimal switching strategy, responding according to either visual or auditory

  15. Concurrent emotional pictures modulate temporal order judgments of spatially separated audio-tactile stimuli.

    PubMed

    Jia, Lina; Shi, Zhuanghua; Zang, Xuelian; Müller, Hermann J

    2013-11-01

    Although attention can be captured toward high-arousal stimuli, little is known about how perceiving emotion in one modality influences the temporal processing of non-emotional stimuli in other modalities. We addressed this issue by presenting observers spatially uninformative emotional pictures while they performed an audio-tactile temporal-order judgment (TOJ) task. In Experiment 1, audio-tactile stimuli were presented at the same location straight ahead of the participants, who had to judge "which modality came first?". In Experiments 2 and 3, the audio-tactile stimuli were delivered one to the left and the other to the right side, and participants had to judge "which side came first?". We found both negative and positive high-arousal pictures to significantly bias TOJs towards the tactile and away from the auditory event when the audio-tactile stimuli were spatially separated; by contrast, there was no such bias when the audio-tactile stimuli originated from the same location. To further examine whether this bias is attributable to the emotional meanings conveyed by the pictures or to their high arousal effect, we compared and contrasted the influences of near-body threat vs. remote threat (emotional) pictures on audio-tactile TOJs in Experiment 3. The bias manifested only in the near-body threat condition. Taken together, the findings indicate that visual stimuli conveying meanings of near-body interaction activate a sensorimotor functional link prioritizing the processing of tactile over auditory signals when these signals are spatially separated. In contrast, audio-tactile signals from the same location engender strong crossmodal integration, thus counteracting modality-based attentional shifts induced by the emotional pictures. PMID:24041776

  16. Detection of emetic activity in the cat by monitoring venous pressure and audio signals

    NASA Technical Reports Server (NTRS)

    Nagahara, A.; Fox, Robert A.; Daunton, Nancy G.; Elfar, S.

    1991-01-01

    To investigate the use of audio signals as a simple, noninvasive measure of emetic activity, the relationship between the somatic events and sounds associated with retching and vomiting was studied. Thoracic venous pressure obtained from an implanted external jugular catheter was shown to provide a precise measure of the somatic events associated with retching and vomiting. Changes in thoracic venous pressure monitored through an indwelling external jugular catheter with audio signals, obtained from a microphone located above the animal in a test chamber, were compared. In addition, two independent observers visually monitored emetic episodes. Retching and vomiting were induced by injection of xylazine (0.66mg/kg s.c.), or by motion. A unique audio signal at a frequency of approximately 250 Hz is produced at the time of the negative thoracic venous pressure change associated with retching. Sounds with higher frequencies (around 2500 Hz) occur in conjunction with the positive pressure changes associated with vomiting. These specific signals could be discriminated reliably by individuals reviewing the audio recordings of the sessions. Retching and those emetic episodes associated with positive venous pressure changes were detected accurately by audio monitoring, with 90 percent of retches and 100 percent of emetic episodes correctly identified. Retching was detected more accurately (p is less than .05) by audio monitoring than by direct visual observation. However, with visual observation a few incidents in which stomach contents were expelled in the absence of positive pressure changes or detectable sounds were identified. These data suggest that in emetic situations, the expulsion of stomach contents may be accomplished by more than one neuromuscular system and that audio signals can be used to detect emetic episodes associated with thoracic venous pressure changes.

  17. Microphone Handling Noise: Measurements of Perceptual Threshold and Effects on Audio Quality

    PubMed Central

    Kendrick, Paul; Jackson, Iain R.; Fazenda, Bruno M.; Cox, Trevor J.; Li, Francis F.

    2015-01-01

    A psychoacoustic experiment was carried out to test the effects of microphone handling noise on perceived audio quality. Handling noise is a problem affecting both amateurs using their smartphones and cameras, as well as professionals using separate microphones and digital recorders. The noises used for the tests were measured from a variety of devices, including smartphones, laptops and handheld microphones. The signal features that characterise these noises are analysed and presented. The sounds include various types of transient, impact noises created by tapping or knocking devices, as well as more sustained sounds caused by rubbing. During the perceptual tests, listeners auditioned speech podcasts and were asked to rate the degradation of any unwanted sounds they heard. A representative design test methodology was developed that tried to encourage everyday rather than analytical listening. Signal-to-noise ratio (SNR) of the handling noise events was shown to be the best predictor of quality degradation. Other factors such as noise type or background noise in the listening environment did not significantly affect quality ratings. Podcast, microphone type and reproduction equipment were found to be significant but only to a small extent. A model allowing the prediction of degradation from the SNR is presented. The SNR threshold at which 50% of subjects noticed handling noise was found to be 4.2 ± 0.6 dBA. The results from this work are important for the understanding of our perception of impact sound and resonant noises in recordings, and will inform the future development of an automated predictor of quality for handling noise. PMID:26473498

  18. Microphone Handling Noise: Measurements of Perceptual Threshold and Effects on Audio Quality.

    PubMed

    Kendrick, Paul; Jackson, Iain R; Fazenda, Bruno M; Cox, Trevor J; Li, Francis F

    2015-01-01

    A psychoacoustic experiment was carried out to test the effects of microphone handling noise on perceived audio quality. Handling noise is a problem affecting both amateurs using their smartphones and cameras, as well as professionals using separate microphones and digital recorders. The noises used for the tests were measured from a variety of devices, including smartphones, laptops and handheld microphones. The signal features that characterise these noises are analysed and presented. The sounds include various types of transient, impact noises created by tapping or knocking devices, as well as more sustained sounds caused by rubbing. During the perceptual tests, listeners auditioned speech podcasts and were asked to rate the degradation of any unwanted sounds they heard. A representative design test methodology was developed that tried to encourage everyday rather than analytical listening. Signal-to-noise ratio (SNR) of the handling noise events was shown to be the best predictor of quality degradation. Other factors such as noise type or background noise in the listening environment did not significantly affect quality ratings. Podcast, microphone type and reproduction equipment were found to be significant but only to a small extent. A model allowing the prediction of degradation from the SNR is presented. The SNR threshold at which 50% of subjects noticed handling noise was found to be 4.2 ± 0.6 dBA. The results from this work are important for the understanding of our perception of impact sound and resonant noises in recordings, and will inform the future development of an automated predictor of quality for handling noise. PMID:26473498

  19. An Audio/Video Reporting Workflow to Supplement Standardized Radiology Reports.

    PubMed

    Balkman, Jason D; Siegel, Alan H

    2016-04-01

    Radiology studies are inherently visual and the information contained within is best conveyed by visual methodology. Advanced reporting software allows the incorporation of annotated key images into text reports, but such features may be less effective compared with in-person consultations. The use of web technology and screen capture software to create retrievable on-demand audio/visual reports has not yet been investigated. This approach may preempt potential curbside consultations while providing referring clinicians with a more engaged imaging service. In this work, we develop and evaluate a video reporting tool that utilizes modern screen capture software and web technology. We hypothesize that referring clinicians would find that recorded on-demand video reports add value to clinical practice, education, and that such technology would be welcome in future practice. A total of 45 case videos were prepared by radiologists for 14 attending and 15 trainee physicians from emergency and internal medicine specialties. Positive survey feedback from referring clinicians about the video reporting system was statistically significant in all areas measured, including video quality, clinical helpfulness, and willingness to use such technology in the future. Trainees unanimously found educational value in video reporting. These results suggest the potential for video technology to re-establish the radiologist's role as a pivotal member of patient care and integral clinical educator. Future work is needed to streamline these methods in order to minimize work redundancy with traditional text reporting. Additionally, integration with an existing PACS and dictation system will be essential to ensuring ease of use and widespread adoption. PMID:26370670

  20. General features

    SciTech Connect

    Wallace, R.E.

    1990-01-01

    The San Andreas fault system, a complex of faults that display predominantly large-scale strike slip, is part of an even more complex system of faults, isolated segments of the East Pacific Rise, and scraps of plates lying east of the East Pacific Rise that collectively separate the North American plate from the Pacific plate. This chapter briefly describes the San Andreas fault system, its setting along the Pacific Ocean margin of North America, its extent, and the patterns of faulting. Only selected characteristics are described, and many features are left for depictions on maps and figures.