Science.gov

Sample records for audio machine text-to-speech

  1. A text-to-speech converter for radiology journal articles.

    PubMed

    Richardson, Michael L

    2010-12-01

    Radiology articles are primarily designed to be read on paper or a screen. Audio versions let users hear this material during activities when reading is not practical. Currently, there are relatively few radiology materials in audio format. However, inexpensive text-to-speech software can easily produce spoken-word versions of digital text. This paper describes a free Web-based program that converts radiology articles to audio format using text-to-speech software. PMID:20863720

  2. Text-to-speech synthesis for Welsh and Welsh English. 

    E-print Network

    Williams, Briony

    1995-01-01

    This work represents the first known attempt to develop a text-to-speech synthesiser for Welsh. A list of pseudo-Welsh nonsense words was generated, allowing for certain difficulties particular to Welsh. Diphones were ...

  3. A markup language for text-to-speech synthesis. 

    E-print Network

    Sproat, Richard; Taylor, Paul; Tanenblatt, Michael; Isard, Amy

    1997-01-01

    Text-to-speech synthesizers must process text, and therefore require some knowledge of text structure. While many TTS systems allow for user control by means of ad hoc ‘escape sequences’, there remains to date no ...

  4. Unsupervised learning for text-to-speech synthesis 

    E-print Network

    Watts, Oliver Samuel

    2013-07-02

    This thesis introduces a general method for incorporating the distributional analysis of textual and linguistic objects into text-to-speech (TTS) conversion systems. Conventional TTS conversion uses intermediate layers ...

  5. The Study and Implementation of Text-to-Speech System for Agricultural Information

    NASA Astrophysics Data System (ADS)

    Zheng, Huoguo; Hu, Haiyan; Liu, Shihong; Meng, Hong

    The Broadcast and Television coverage has increased to more than 98% in china. Information services by radio have wide coverage, low cost, easy-to-grass-roots farmers to accept etc. characteristics. In order to play the better role of broadcast information service, as well as aim at the problem of lack of information resource in rural, we R & D the text-to-speech system. The system includes two parts, software and hardware device, both of them can translate text into audio file. The software subsystem was implemented basic on third-part middleware, and the hardware subsystem was realized with microelectronics technology. Results indicate that the hardware is better than software. The system has been applied in huailai city hebei province, which has conversed more than 8000 audio files as programming materials for the local radio station.

  6. "Look What I Did!": Student Conferences with Text-to-Speech Software

    ERIC Educational Resources Information Center

    Young, Chase; Stover, Katie

    2014-01-01

    The authors describe a strategy that empowers students to edit and revise their own writing. Students input their writing in to text-to-speech software that rereads the text aloud. While listening, students make necessary revisions and edits.

  7. Using prosodic structure to improve pitch range variation in text to speech synthesis. 

    E-print Network

    Clark, Robert A J

    1999-01-01

    The intonation produced by current text-to-speech systems is often either flat or artificial sounding. Pitch range is one of the contributing factors which could be improved by more detailed linguistic knowledge. In ...

  8. Concatenative Text-to-Speech Synthesis Based on Prototype Waveform Interpolation (A Time Frequency Approach) 

    E-print Network

    Morais, Edmilson; Taylor, Paul; Violaro, Fabio

    This paper presents some preliminary methods to apply the Time- Frequency Interpolation technique - TFI [3] to concatenative text-to-speech synthesis. The TFI technique described here is a pitch-synchronous time-frequency ...

  9. Identifying prosodic prominence patterns for English text-to-speech synthesis 

    E-print Network

    Badino, Leonardo

    2010-01-01

    This thesis proposes to improve and enrich the expressiveness of English Text-to-Speech (TTS) synthesis by identifying and generating natural patterns of prosodic prominence. In most state-of-the-art TTS systems the ...

  10. Including Pitch Accent Optionality in Unit Selection Text-to-Speech Synthesis 

    E-print Network

    Badino, Leonardo; Robert A J, Clark; Strom, Volker

    2008-01-01

    this variability to synthesize prosodic prominence in unit selection text-to-speech synthesis. The main motivation of our approach is that by taking into account the variability of accent placements we enlarge the set of prosodically acceptable speech units, thus...

  11. Integrating Text-to-Speech Software into Pedagogically Sound Teaching and Learning Scenarios

    ERIC Educational Resources Information Center

    Rughooputh, S. D. D. V.; Santally, M. I.

    2009-01-01

    This paper presents a new technique of delivery of classes--an instructional technique which will no doubt revolutionize the teaching and learning, whether for on-campus, blended or online modules. This is based on the simple task of instructionally incorporating text-to-speech software embedded in the lecture slides that will simulate exactly the…

  12. MORPHOLOGICAL ANALYSIS FOR A GERMAN TEXT-TO-SPEECH SYSTEM Amanda Pounder, Markus Kommenda

    E-print Network

    MORPHOLOGICAL ANALYSIS FOR A GERMAN TEXT-TO-SPEECH SYSTEM Amanda Pounder, Markus Kommenda Institut is the automatic derivation of correct pronunciation from the graphemic form of a text. The software module GRAPHON for German word-forms. It provides each text input item with an individual characterization

  13. Advancements in text-to-speech technology and implications for AAC applications

    NASA Astrophysics Data System (ADS)

    Syrdal, Ann K.

    2003-10-01

    Intelligibility was the initial focus in text-to-speech (TTS) research, since it is clearly a necessary condition for the application of the technology. Sufficiently high intelligibility (approximating human speech) has been achieved in the last decade by the better formant-based and concatenative TTS systems. This led to commercially available TTS systems for highly motivated users, particularly the blind and vocally impaired. Some unnatural qualities of TTS were exploited by these users, such as very fast speaking rates and altered pitch ranges for flagging relevant information. Recently, the focus in TTS research has turned to improving naturalness, so that synthetic speech sounds more human and less robotic. Unit selection approaches to concatenative synthesis have dramatically improved TTS quality, although at the cost of larger and more complex systems. This advancement in naturalness has made TTS technology more acceptable to the general public. The vocally impaired appreciate a more natural voice with which to represent themselves when communicating with others. Unit selection TTS does not achieve such high speaking rates as the earlier TTS systems, however, which is a disadvantage to some AAC device users. An important new research emphasis is to improve and increase the range of emotional expressiveness of TTS.

  14. Developing a Test Bed of English Text-to-Speech System XIMERA for the Blizzard Challenge 2006

    E-print Network

    Duh, Kevin

    Developing a Test Bed of English Text-to-Speech System XIMERA for the Blizzard Challenge 2006-Speech (TTS) system XIMERA at ATR for Blizzard Challenge 2006. The original XIMERA is aimed at constructing speech corpora. In order to participate in the Blizzard Challenge 2006, we construct a test bed

  15. Emerging Realities of Text-to-Speech Software for Nonnative-English-Speaking Community College Students in the Freshman Year

    ERIC Educational Resources Information Center

    Baker, Fiona S.

    2015-01-01

    This study explores the expectations and early and subsequent realities of text-to-speech software for 24 nonnative-English-speaking college students who were experiencing reading difficulties in their freshman year of college. The study took place over two semesters in one academic year (from September to June) at a community college on the…

  16. Supporting Reading Comprehension of At-Risk Pre-Adolescent Readers through the Use of Text-to-Speech Technology Paired with Strategic Instruction

    ERIC Educational Resources Information Center

    Anderson, Susan D.

    2009-01-01

    This research highlighted the use of text-to-speech technology and current shifts in strategy-based reading instruction in order to address the comprehension needs of struggling pre-adolescent readers. The following questions were posed: (a) Does reading comprehension of preadolescent struggling readers improve as the direct result of using…

  17. Audio 2008: Audio Fixation

    ERIC Educational Resources Information Center

    Kaye, Alan L.

    2008-01-01

    Take a look around the bus or subway and see just how many people are bumping along to an iPod or an MP3 player. What they are listening to is their secret, but the many signature earbuds in sight should give one a real sense of just how pervasive digital audio has become. This article describes how that popularity is mirrored in library audio

  18. Targeted Audio

    NASA Astrophysics Data System (ADS)

    Olszewski, Dirk

    Targeted audio aims at creating personal listening zones by utilizing adequate measurements. A person inside this listening zone shall be able to perceive acoustically submitted information without disturbing other persons outside the desired listening zone. In order to fulfill this demand, the use of a highly directional audible sound beam is favored. The sound beam shall be aimed at the respective listening zone target, thus implicating the expression targeted audio.

  19. Fast transcription of unstructured audio recordings

    E-print Network

    Roy, Brandon Cain

    We introduce a new method for human-machine collaborative speech transcription that is significantly faster than existing transcription methods. In this approach, automatic audio processing algorithms are used to robustly ...

  20. 2005 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics October 16-19, 2005, New Paltz, NY LEARNING AUDITORY MODELS OF MACHINE VOICES

    E-print Network

    Ellis, Dan

    to traditional therapy akin to art therapy and music therapy, utilizes the sounds of machines as relational@ee.columbia.edu ABSTRACT Vocal imitation is often found useful in Machine Therapy ses- sions as it creates an emphatic by our work in Machine Therapy in which hu- mans try to vocally imitate machines, but this task also

  1. Detecting double compression of audio signal

    NASA Astrophysics Data System (ADS)

    Yang, Rui; Shi, Yun Q.; Huang, Jiwu

    2010-01-01

    MP3 is the most popular audio format nowadays in our daily life, for example music downloaded from the Internet and file saved in the digital recorder are often in MP3 format. However, low bitrate MP3s are often transcoded to high bitrate since high bitrate ones are of high commercial value. Also audio recording in digital recorder can be doctored easily by pervasive audio editing software. This paper presents two methods for the detection of double MP3 compression. The methods are essential for finding out fake-quality MP3 and audio forensics. The proposed methods use support vector machine classifiers with feature vectors formed by the distributions of the first digits of the quantized MDCT (modified discrete cosine transform) coefficients. Extensive experiments demonstrate the effectiveness of the proposed methods. To the best of our knowledge, this piece of work is the first one to detect double compression of audio signal.

  2. Video salient event classification using audio features

    NASA Astrophysics Data System (ADS)

    Corchs, Silvia; Ciocca, Gianluigi; Fiori, Massimiliano; Gasparini, Francesca

    2014-03-01

    The aim of this work is to detect the events in video sequences that are salient with respect to the audio signal. In particular, we focus on the audio analysis of a video, with the goal of finding which are the significant features to detect audio-salient events. In our work we have extracted the audio tracks from videos of different sport events. For each video, we have manually labeled the salient audio-events using the binary markings. On each frame, features in both time and frequency domains have been considered. These features have been used to train different classifiers: Classification and Regression Trees, Support Vector Machine, and k-Nearest Neighbor. The classification performances are reported in terms of confusion matrices.

  3. Audio-visual affective expression recognition

    NASA Astrophysics Data System (ADS)

    Huang, Thomas S.; Zeng, Zhihong

    2007-11-01

    Automatic affective expression recognition has attracted more and more attention of researchers from different disciplines, which will significantly contribute to a new paradigm for human computer interaction (affect-sensitive interfaces, socially intelligent environments) and advance the research in the affect-related fields including psychology, psychiatry, and education. Multimodal information integration is a process that enables human to assess affective states robustly and flexibly. In order to understand the richness and subtleness of human emotion behavior, the computer should be able to integrate information from multiple sensors. We introduce in this paper our efforts toward machine understanding of audio-visual affective behavior, based on both deliberate and spontaneous displays. Some promising methods are presented to integrate information from both audio and visual modalities. Our experiments show the advantage of audio-visual fusion in affective expression recognition over audio-only or visual-only approaches.

  4. Perceptually Based Audio Coding

    NASA Astrophysics Data System (ADS)

    Houtsma, Adrianus J. M.

    High-quality audio is a concept that is not exactly defined and not always properly understood. To some, it refers directly to the physical similarity between a real sound field and its electroacoustical reproduction. In this viewpoint, acoustical knowledge and electronic technology are the only limiting factors preventing audio quality from being perfect. To others, however, audio quality refers to the audible similarity between a real life sound event and an electronic reproduction. Given this viewpoint, the human auditory system with all its limitations becomes an essential factor determining audio quality.

  5. Acoustic chase : designing an interactive audio environment to stimulate human body movement

    E-print Network

    Schiessl, Simon Karl Josef, 1972-

    2004-01-01

    An immersive audio environment was created that explores how humans react to commands imposed by a machine generating its acoustic stimuli on the basis of tracked body movement. In this environment, different states of ...

  6. PERMUTATION GROUPING: INTELLIGENT HASH FUNCTION DESIGN FOR AUDIO & IMAGE RETRIEVAL

    E-print Network

    Cortes, Corinna

    PERMUTATION GROUPING: INTELLIGENT HASH FUNCTION DESIGN FOR AUDIO & IMAGE RETRIEVAL Shumeet Baluja systems. In this study, we introduce the idea of permutation-grouping to intelligently design the hash for retrieval [1]. It was designed to hold 108 -109 keys, distributed to a network of machines. Despite

  7. Audio Engineering Society Convention Paper

    E-print Network

    Jackson, Philip JB

    in the spatial quality of reproduced audio, rather than changes in the timbral quality. This contrasts with previous models of audio quality, such as PEAQ (ITU-R BS1387) [16], which have not explicitly con- sideredAudio Engineering Society Convention Paper Presented at the 125th Convention 2008 October 2­5 San

  8. Audio Engineering Society Convention Paper

    E-print Network

    Jackson, Philip JB

    .rumsey@surrey.ac.uk ABSTRACT Most current perceptual models for audio quality have so far tended to concentrate those obtained in listening tests. 1. INTRODUCTION Most current perceptual models for audio quality have the audibility of codec distortions in terms of basic audio quality, or mean opinion score, for example, and have

  9. Real World Audio

    NASA Technical Reports Server (NTRS)

    1998-01-01

    Crystal River Engineering was originally featured in Spinoff 1992 with the Convolvotron, a high speed digital audio processing system that delivers three-dimensional sound over headphones. The Convolvotron was developed for Ames' research on virtual acoustic displays. Crystal River is a now a subsidiary of Aureal Semiconductor, Inc. and they together develop and market the technology, which is a 3-D (three dimensional) audio technology known commercially today as Aureal 3D (A-3D). The technology has been incorporated into video games, surround sound systems, and sound cards.

  10. 3D Audio System

    NASA Technical Reports Server (NTRS)

    1992-01-01

    Ames Research Center research into virtual reality led to the development of the Convolvotron, a high speed digital audio processing system that delivers three-dimensional sound over headphones. It consists of a two-card set designed for use with a personal computer. The Convolvotron's primary application is presentation of 3D audio signals over headphones. Four independent sound sources are filtered with large time-varying filters that compensate for motion. The perceived location of the sound remains constant. Possible applications are in air traffic control towers or airplane cockpits, hearing and perception research and virtual reality development.

  11. Efficient audio signal processing for embedded systems

    NASA Astrophysics Data System (ADS)

    Chiu, Leung Kin

    As mobile platforms continue to pack on more computational power, electronics manufacturers start to differentiate their products by enhancing the audio features. However, consumers also demand smaller devices that could operate for longer time, hence imposing design constraints. In this research, we investigate two design strategies that would allow us to efficiently process audio signals on embedded systems such as mobile phones and portable electronics. In the first strategy, we exploit properties of the human auditory system to process audio signals. We designed a sound enhancement algorithm to make piezoelectric loudspeakers sound ”richer" and "fuller." Piezoelectric speakers have a small form factor but exhibit poor response in the low-frequency region. In the algorithm, we combine psychoacoustic bass extension and dynamic range compression to improve the perceived bass coming out from the tiny speakers. We also developed an audio energy reduction algorithm for loudspeaker power management. The perceptually transparent algorithm extends the battery life of mobile devices and prevents thermal damage in speakers. This method is similar to audio compression algorithms, which encode audio signals in such a ways that the compression artifacts are not easily perceivable. Instead of reducing the storage space, however, we suppress the audio contents that are below the hearing threshold, therefore reducing the signal energy. In the second strategy, we use low-power analog circuits to process the signal before digitizing it. We designed an analog front-end for sound detection and implemented it on a field programmable analog array (FPAA). The system is an example of an analog-to-information converter. The sound classifier front-end can be used in a wide range of applications because programmable floating-gate transistors are employed to store classifier weights. Moreover, we incorporated a feature selection algorithm to simplify the analog front-end. A machine learning algorithm AdaBoost is used to select the most relevant features for a particular sound detection application. In this classifier architecture, we combine simple "base" analog classifiers to form a strong one. We also designed the circuits to implement the AdaBoost-based analog classifier.

  12. Audio Feedback -- Better Feedback?

    ERIC Educational Resources Information Center

    Voelkel, Susanne; Mello, Luciane V.

    2014-01-01

    National Student Survey (NSS) results show that many students are dissatisfied with the amount and quality of feedback they get for their work. This study reports on two case studies in which we tried to address these issues by introducing audio feedback to one undergraduate (UG) and one postgraduate (PG) class, respectively. In case study one…

  13. (Foto, Audio, Video) Multimediln prvky

    E-print Network

    Slovák, Jan

    (Foto, Audio, Video) Multimediální prvky v Odpovdníku IS MU Servisní stedisko pro podporu e odpovdník? 4 Foto Vlození obrázku, fotografie do Odpovdníku IS MU 6 1. Mám obrázek, ale potebuji zmnit jeho multimediálních prvk v Odpovdnících IS MU. Ve tech hlavních kapitolách ­ Foto, Audio, Video ­ se nejprve seznámíte

  14. The Lowdown on Audio Downloads

    ERIC Educational Resources Information Center

    Farrell, Beth

    2010-01-01

    First offered to public libraries in 2004, downloadable audiobooks have grown by leaps and bounds. According to the Audio Publishers Association, their sales today account for 21% of the spoken-word audio market. It hasn't been easy, however. WMA. DRM. MP3. AAC. File extensions small on letters but very big on consequences for librarians,…

  15. Engaging Students with Audio Feedback

    ERIC Educational Resources Information Center

    Cann, Alan

    2014-01-01

    Students express widespread dissatisfaction with academic feedback. Teaching staff perceive a frequent lack of student engagement with written feedback, much of which goes uncollected or unread. Published evidence shows that audio feedback is highly acceptable to students but is underused. This paper explores methods to produce and deliver audio

  16. Audio issues in MIR evaluation vOverview of audio formats

    E-print Network

    Reiss, Josh

    regardless of internal data #12;Simple Guidelines #1 · Audio files should be presented in the highest quality? · Guarantee highest quality ­ Then test robustness · Far richer data ü20+ tracks ü96kHz + ü24 bit · Many audioAudio issues in MIR evaluation vOverview of audio formats vPreferred presentation of audio files

  17. Audio Engineering Society Convention Paper 7599

    E-print Network

    Jackson, Philip JB

    of sound quality assessment. The existing objective models for predicting quality scores of broadband audio audio quality, new features that describe spatial characteristics of surround sound have that predicts Basic Audio Quality (BAQ) of multichannel audio recordings encoded by perceptual encoders. However

  18. Audio browsing of automaton-based hypertext 

    E-print Network

    Ustun, Selen

    2004-09-30

    is that the proposed Audio Browser is an efficient tool for presenting hypertext in audio format, which will prove to be useful for several applications including browsers for visually-impaired and remote users. The Audio Browser provides audio-only browsing...

  19. Digital Audio Compression By Davis Yen Pan

    E-print Network

    Ellis, Dan

    Digital Audio Compression By Davis Yen Pan Abstract Compared to most digital data types, with the exception of digital video, the data rates associ- ated with uncompressed digital audio are substan- tial. Digital audio compression enables more effi- cient storage and transmission of audio data. The many forms

  20. Metrological digital audio reconstruction

    DOEpatents

    Fadeyev; Vitaliy (Berkeley, CA), Haber; Carl (Berkeley, CA)

    2004-02-19

    Audio information stored in the undulations of grooves in a medium such as a phonograph record may be reconstructed, with little or no contact, by measuring the groove shape using precision metrology methods coupled with digital image processing and numerical analysis. The effects of damage, wear, and contamination may be compensated, in many cases, through image processing and analysis methods. The speed and data handling capacity of available computing hardware make this approach practical. Two examples used a general purpose optical metrology system to study a 50 year old 78 r.p.m. phonograph record and a commercial confocal scanning probe to study a 1920's celluloid Edison cylinder. Comparisons are presented with stylus playback of the samples and with a digitally re-mastered version of an original magnetic recording. There is also a more extensive implementation of this approach, with dedicated hardware and software.

  1. The Audio-Tutorial System

    ERIC Educational Resources Information Center

    Postlethwait, S. N.

    1970-01-01

    Describes the audio-tutorial program in Botany at Purdue University. Advantages include adaptability to individual stduent needs, integration of laboratory activities and information giving, aid flexibility in use of media and means of presentation. (EB)

  2. Plasmon-Assisted Audio Recording

    E-print Network

    Chen, Hao

    We present the first demonstration of the recording of optically encoded audio onto a plasmonic nanostructure. Analogous to the ‘‘optical sound’’ approach used in the early twentieth century to store sound on photographic ...

  3. A centralized audio presentation manager

    SciTech Connect

    Papp, A.L. III; Blattner, M.M.

    1994-05-16

    The centralized audio presentation manager addresses the problems which occur when multiple programs running simultaneously attempt to use the audio output of a computer system. Time dependence of sound means that certain auditory messages must be scheduled simultaneously, which can lead to perceptual problems due to psychoacoustic phenomena. Furthermore, the combination of speech and nonspeech audio is examined; each presents its own problems of perceptibility in an acoustic environment composed of multiple auditory streams. The centralized audio presentation manager receives abstract parameterized message requests from the currently running programs, and attempts to create and present a sonic representation in the most perceptible manner through the use of a theoretically and empirically designed rule set.

  4. Robot Command Interface Using an Audio-Visual Speech Recognition System

    NASA Astrophysics Data System (ADS)

    Ceballos, Alexánder; Gómez, Juan; Prieto, Flavio; Redarce, Tanneguy

    In recent years audio-visual speech recognition has emerged as an active field of research thanks to advances in pattern recognition, signal processing and machine vision. Its ultimate goal is to allow human-computer communication using voice, taking into account the visual information contained in the audio-visual speech signal. This document presents a command's automatic recognition system using audio-visual information. The system is expected to control the laparoscopic robot da Vinci. The audio signal is treated using the Mel Frequency Cepstral Coefficients parametrization method. Besides, features based on the points that define the mouth's outer contour according to the MPEG-4 standard are used in order to extract the visual speech information.

  5. AUDIO FINGERPRINTING: COMBINING COMPUTER VISION & DATA STREAM PROCESSING

    E-print Network

    Cortes, Corinna

    AUDIO FINGERPRINTING: COMBINING COMPUTER VISION & DATA STREAM PROCESSING Shumeet Baluja & Michele fingerprints of audio data that can be efficiently matched. The resulting system has excellent identification. INTRODUCTION & BACKGROUND Audio fingerprinting provides the ability to link short, unlabeled, snippets of audio

  6. Text to Speech: A 4-H Model of Accessibility and Inclusion

    ERIC Educational Resources Information Center

    Green, Jeremy W.

    2012-01-01

    4-H project manuals play an integral part in a youth's ability to achieve mastery in a specific project area. For youth who struggle with reading, written 4-H materials prove inadequate in addressing the needs of the learner. This article proposes a new delivery method of 4-H educational material designed to create a more inclusive and…

  7. TEXT-TO-SPEECH SYNTHESIS FOR WELSH AND WELSH ENGLISH Briony Williams

    E-print Network

    Edinburgh, University of

    Research, University of Edinburgh, 80 South Bridge Edinburgh EH1 1HN, Scotland, UK email: briony extra segments were added to cover HG6 HG6 FI6 - C A ' F D A - 8 A - BPA EGA H H A - ¡£¢ - ¤¥¢ Table 2

  8. Expressive Visual Text-To-Speech as an Assistive Technology for Individuals with Autism Spectrum Conditions

    E-print Network

    Cassidy, S. A.; Stenger, B.; Van Dongen, L.; Yanagisawa, K.; Anderson, R.; Wan, V.; Baron-Cohen, S.; Cipolla, R.

    2015-01-01

    Adults with Autism Spectrum Conditions (ASC) experience marked difficulties in recognizing the emotions of others and responding appropriately. The clinical characteristics of ASC mean that face to face or group interventions may not be appropriate...

  9. Robust Speaker-Adaptive HMM-based Text-to-Speech Synthesis 

    E-print Network

    Yamagishi, Junichi; Nose, Takashi; Zen, Heiga; Ling, Zhenhua; Toda, Tomoki; Tokuda, Keiichi; King, Simon; Renals, Steve

    2009-01-01

    This paper describes a speaker-adaptive HMM-based speech synthesis system. The new system, called ``HTS-2007,'' employs speaker adaptation (CSMAPLR+MAP), feature-space adaptive training, mixed-gender modeling, and full-covariance modeling using...

  10. Bayesian modelling of vowel segment duration for text-to-speech synthesis using distinctive features 

    E-print Network

    Goubanova, Olga V

    2003-01-01

    also experimented with a word class feature that implicitly represents word frequency information. We contrasted the results of the belief network model with those of the sums of products (SoP) model and classification and regression tree (CART) model...

  11. Computational Linguistics Volume 24, Number 2 An Introduction to Text-to-Speech Synthesis

    E-print Network

    Dutoit (Facult4 Polytechnique de Mons) Dordrecht: Kluwer Academic Publishers (Text, Speech and Language divided between the natural language processing (NLP) components of a TTS system and the digital signal and weaknesses, is a real gift to the reader who is using this book to make design choices. Because the book

  12. REISS ET AL. COMPRESSION FOR SUPER AUDIO CD Audio Engineering Society

    E-print Network

    Reiss, Josh

    been to audio quality and additional functionality. Furthermore, DSD and related sigma delta modulationREISS ET AL. COMPRESSION FOR SUPER AUDIO CD _________________________________ Audio Engineering be obtained by sending request and remittance to Audio Engineering Society, 60 East 42nd Street, New York, New

  13. Low Bitrate Coding of Spot Audio Signals for Interactive and Immersive Audio Applications

    E-print Network

    Tsakalides, Panagiotis

    between the user and the environment. A model is proposed which achieves high-quality audio re- productionLow Bitrate Coding of Spot Audio Signals for Interactive and Immersive Audio Applications years, a revolution has occurred in the area of consumer audio. Similarly to the transition from analog

  14. Low Bitrate Coding of Spot Audio Signals for Interactive and Immersive Audio Applications

    E-print Network

    Tsakalides, Panagiotis

    between the user and the environment. A model is proposed which achieves high-quality audio reproductionLow Bitrate Coding of Spot Audio Signals for Interactive and Immersive Audio Applications, a revolution has occurred in the area of consumer audio. Similarly to the transition from analog to digital

  15. Audio classification from time-frequency texture

    E-print Network

    Slotine, Jean-Jacques E.

    Time-frequency representations of audio signals often resemble texture images. This paper derives a simple audio classification algorithm based on treating sound spectrograms as texture images. The algorithm is inspired ...

  16. Timbre-invariant Audio Features for Style Analysis of Classical Music Christof Wei

    E-print Network

    Mauch, Matthias

    Timbre-invariant Audio Features for Style Analysis of Classical Music Christof Weiß Fraunhofer IDMT, more recently, automatic analyses on symbolic data, both largely independent from timbre. Here, we in that is invariant to timbre. Classification using Support Vec- tor Machines shows that a high era classification

  17. DRAFT, TO APPEAR IN TRANSACTIONS ON SPEECH AND AUDIO PROCESSING 1 Speaker verification using

    E-print Network

    Edinburgh, University of

    DRAFT, TO APPEAR IN TRANSACTIONS ON SPEECH AND AUDIO PROCESSING 1 Speaker verification using a text-independent speaker verification system using support vector machines (SVMs) with score such as Gaussian mixture models (GMMs). This approach provides direct discrim- ination between whole sequences

  18. DRAFT, TO APPEAR IN TRANSACTIONS ON SPEECH AND AUDIO PROCESSING 1 Speaker verification using

    E-print Network

    Edinburgh, University of

    DRAFT, TO APPEAR IN TRANSACTIONS ON SPEECH AND AUDIO PROCESSING 1 Speaker verification using a text­independent speaker verification system using support vector machines (SVMs) with score such as Gaussian mixture models (GMMs). This approach provides direct discrim­ ination between whole sequences

  19. Conetta et al Audio Engineering Society

    E-print Network

    Jackson, Philip JB

    the reference and impaired items' in an audio system [ITU-R BS.1534, 2001], `spatial quality' is defined hereConetta et al Audio Engineering Society Convention Paper Presented at the 125th Convention 2008. Additional papers may be obtained by sending request and remittance to Audio Engineering Society, 60 East 42

  20. Spatialized Audio Rendering for Immersive Virtual Environments

    E-print Network

    Staadt, Oliver

    trade-off between audio quality, spatial precision, and perfor- mance. Convincing acoustic roomSpatialized Audio Rendering for Immersive Virtual Environments Martin Naef1 Oliver Staadt2 Markus@cs.ucdavis.edu ABSTRACT We present a spatialized audio rendering system for the use in immersive virtual environments

  1. AUTOMATIC TRANSCRIPTION OF COMPRESSED BROADCAST AUDIO

    E-print Network

    are compared with those obtained on high quality PCM recordings prior to compres- sion. For a 6.5 kbps audio decrease in the transcription quality, and how to improve the system for audio at higher compression ratesAUTOMATIC TRANSCRIPTION OF COMPRESSED BROADCAST AUDIO Claude Barras, Lori Lamel and Jean

  2. Audio-Visual Aids: Historians in Blunderland.

    ERIC Educational Resources Information Center

    Decarie, Graeme

    1988-01-01

    A history professor relates his experiences producing and using audio-visual material and warns teachers not to rely on audio-visual aids for classroom presentations. Includes examples of popular audio-visual aids on Canada that communicate unintended, inaccurate, or unclear ideas. Urges teachers to exercise caution in the selection and use of…

  3. Audio-visual Materials and Rural Libraries

    ERIC Educational Resources Information Center

    Escolar-Sobrino, Hipolito

    1972-01-01

    Audio-visual materials enlarge the educational work being done in the classroom and the library. This article examines the various types of audio-visual material and equipment and suggests ways in which audio-visual media can be used economically and efficiently in rural libraries. (Author)

  4. Advances in audio source seperation and multisource audio content retrieval

    NASA Astrophysics Data System (ADS)

    Vincent, Emmanuel

    2012-06-01

    Audio source separation aims to extract the signals of individual sound sources from a given recording. In this paper, we review three recent advances which improve the robustness of source separation in real-world challenging scenarios and enable its use for multisource content retrieval tasks, such as automatic speech recognition (ASR) or acoustic event detection (AED) in noisy environments. We present a Flexible Audio Source Separation Toolkit (FASST) and discuss its advantages compared to earlier approaches such as independent component analysis (ICA) and sparse component analysis (SCA). We explain how cues as diverse as harmonicity, spectral envelope, temporal fine structure or spatial location can be jointly exploited by this toolkit. We subsequently present the uncertainty decoding (UD) framework for the integration of audio source separation and audio content retrieval. We show how the uncertainty about the separated source signals can be accurately estimated and propagated to the features. Finally, we explain how this uncertainty can be efficiently exploited by a classifier, both at the training and the decoding stage. We illustrate the resulting performance improvements in terms of speech separation quality and speaker recognition accuracy.

  5. Audio-Visual Materials Catalog.

    ERIC Educational Resources Information Center

    Anderson (M.D.) Hospital and Tumor Inst., Houston, TX.

    This catalog lists 27 audiovisual programs produced by the Department of Medical Communications of the University of Texas M. D. Anderson Hospital and Tumor Institute for public distribution. Video tapes, 16 mm. motion pictures and slide/audio series are presented dealing mostly with cancer and related subjects. The programs are intended for…

  6. Audio/ Videoconferencing Packages: Low Cost

    ERIC Educational Resources Information Center

    Treblay, Remy; Fyvie, Barb; Koritko, Brenda

    2005-01-01

    A comparison was conducted of "Voxwire MeetingRoom" and "iVocalize" v4.1.0.3, both Web-conferencing products using voice-over-Internet protocol (VoIP) to provide unlimited, inexpensive, international audio communication, and high-quality Web-conferencing fostering collaborative learning. The study used the evaluation criteria used in earlier…

  7. Audio Engineering Society Convention Paper

    E-print Network

    Reiss, Josh

    portion thereof, is not permitted without direct permission from the Journal of the Audio Engineering sources around the speaker array. The second objective and main focus of the paper is localising sources within the speaker array. We seek to determine if the sound sources can be located without movement

  8. Radioactive Decay: Audio Data Collection

    ERIC Educational Resources Information Center

    Struthers, Allan

    2009-01-01

    Many phenomena generate interesting audible time series. This data can be collected and processed using audio software. The free software package "Audacity" is used to demonstrate the process by recording, processing, and extracting click times from an inexpensive radiation detector. The high quality of the data is demonstrated with a simple…

  9. A Simple Audio Conductivity Device.

    ERIC Educational Resources Information Center

    Berenato, Gregory; Maynard, David F.

    1997-01-01

    Describes a simple audio conductivity device built to address the problem of the lack of sensitivity needed to measure small differences in conductivity in crude conductivity devices. Uses a 9-V battery as a power supply and allows the relative resistance differences between substances to be detected by the frequency of its audible tones. Presents…

  10. Audio Engineering Society Convention Paper

    E-print Network

    Reiss, Josh

    , construction and analysis of an interactive gesture-controlled audio mixing system by means of a wireless video game controller. The concept is based on the idea that the mixing engineer can step away from and results are reported. 1. INTRODUCTION There are many benefits of a wireless gesture controlled mixing

  11. Quantitative characterisation of audio data by ordinal symbolic dynamics

    NASA Astrophysics Data System (ADS)

    Aschenbrenner, T.; Monetti, R.; Amigó, J. M.; Bunk, W.

    2013-06-01

    Ordinal symbolic dynamics has developed into a valuable method to describe complex systems. Recently, using the concept of transcripts, the coupling behaviour of systems was assessed, combining the properties of the symmetric group with information theoretic ideas. In this contribution, methods from the field of ordinal symbolic dynamics are applied to the characterisation of audio data. Coupling complexity between frequency bands of solo violin music, as a fingerprint of the instrument, is used for classification purposes within a support vector machine scheme. Our results suggest that coupling complexity is able to capture essential characteristics, sufficient to distinguish among different violins.

  12. Development of Learning Modules for Machine Shop Occupations. Final Report.

    ERIC Educational Resources Information Center

    Kent, Randall

    This final report contains an eight-page narrative and materials/products of a program to produce (the final) sixty-eight individualized machine shop skill tasks modules (and fifty-two master audio tapes for students with serious reading disabilities). The narrative also describes the determination of the vital few skills used by machine tool…

  13. Cluster: Metals. Course: Machine Shop. Research Project.

    ERIC Educational Resources Information Center

    Sanford - Lee County Schools, NC.

    The set of 13 units is designed for use with an instructor in actual machine shop practice and is also keyed to audio visual and textual materials. Each unit contains a series of task packages which: specify prerequisites within the series (minimum is Unit 1); provide a narrative rationale for learning; list both general and specific objectives in…

  14. pyAudioAnalysis: An Open-Source Python Library for Audio Signal Analysis

    PubMed Central

    Giannakopoulos, Theodoros

    2015-01-01

    Audio information plays a rather important role in the increasing digital content that is available today, resulting in a need for methodologies that automatically analyze such content: audio event recognition for home automations and surveillance systems, speech recognition, music information retrieval, multimodal analysis (e.g. audio-visual analysis of online videos for content-based recommendation), etc. This paper presents pyAudioAnalysis, an open-source Python library that provides a wide range of audio analysis procedures including: feature extraction, classification of audio signals, supervised and unsupervised segmentation and content visualization. pyAudioAnalysis is licensed under the Apache License and is available at GitHub (https://github.com/tyiannak/pyAudioAnalysis/). Here we present the theoretical background behind the wide range of the implemented methodologies, along with evaluation metrics for some of the methods. pyAudioAnalysis has been already used in several audio analysis research applications: smart-home functionalities through audio event detection, speech emotion recognition, depression classification based on audio-visual features, music segmentation, multimodal content-based movie recommendation and health applications (e.g. monitoring eating habits). The feedback provided from all these particular audio applications has led to practical enhancement of the library. PMID:26656189

  15. pyAudioAnalysis: An Open-Source Python Library for Audio Signal Analysis.

    PubMed

    Giannakopoulos, Theodoros

    2015-01-01

    Audio information plays a rather important role in the increasing digital content that is available today, resulting in a need for methodologies that automatically analyze such content: audio event recognition for home automations and surveillance systems, speech recognition, music information retrieval, multimodal analysis (e.g. audio-visual analysis of online videos for content-based recommendation), etc. This paper presents pyAudioAnalysis, an open-source Python library that provides a wide range of audio analysis procedures including: feature extraction, classification of audio signals, supervised and unsupervised segmentation and content visualization. pyAudioAnalysis is licensed under the Apache License and is available at GitHub (https://github.com/tyiannak/pyAudioAnalysis/). Here we present the theoretical background behind the wide range of the implemented methodologies, along with evaluation metrics for some of the methods. pyAudioAnalysis has been already used in several audio analysis research applications: smart-home functionalities through audio event detection, speech emotion recognition, depression classification based on audio-visual features, music segmentation, multimodal content-based movie recommendation and health applications (e.g. monitoring eating habits). The feedback provided from all these particular audio applications has led to practical enhancement of the library. PMID:26656189

  16. Analysis of musical expression in audio signals

    NASA Astrophysics Data System (ADS)

    Dixon, Simon

    2003-01-01

    In western art music, composers communicate their work to performers via a standard notation which specificies the musical pitches and relative timings of notes. This notation may also include some higher level information such as variations in the dynamics, tempo and timing. Famous performers are characterised by their expressive interpretation, the ability to convey structural and emotive information within the given framework. The majority of work on audio content analysis focusses on retrieving score-level information; this paper reports on the extraction of parameters describing the performance, a task which requires a much higher degree of accuracy. Two systems are presented: BeatRoot, an off-line beat tracking system which finds the times of musical beats and tracks changes in tempo throughout a performance, and the Performance Worm, a system which provides a real-time visualisation of the two most important expressive dimensions, tempo and dynamics. Both of these systems are being used to process data for a large-scale study of musical expression in classical and romantic piano performance, which uses artificial intelligence (machine learning) techniques to discover fundamental patterns or principles governing expressive performance.

  17. Aeronautical audio broadcasting via satellite

    NASA Technical Reports Server (NTRS)

    Tzeng, Forrest F.

    1993-01-01

    A system design for aeronautical audio broadcasting, with C-band uplink and L-band downlink, via Inmarsat space segments is presented. Near-transparent-quality compression of 5-kHz bandwidth audio at 20.5 kbit/s is achieved based on a hybrid technique employing linear predictive modeling and transform-domain residual quantization. Concatenated Reed-Solomon/convolutional codes with quadrature phase shift keying are selected for bandwidth and power efficiency. RF bandwidth at 25 kHz per channel, and a decoded bit error rate at 10(exp -6) with E(sub b)/N(sub o) at 3.75 dB are obtained. An interleaver, scrambler, modem synchronization, and frame format were designed, and frequency-division multiple access was selected over code-division multiple access. A link budget computation based on a worst-case scenario indicates sufficient system power margins. Transponder occupancy analysis for 72 audio channels demonstrates ample remaining capacity to accommodate emerging aeronautical services.

  18. DRAFT FOR IEEE TRANS. ON SPEECH AND AUDIO PROCESSING 1 An experimental comparison of audio tempo

    E-print Network

    Tzanetakis, George

    DRAFT FOR IEEE TRANS. ON SPEECH AND AUDIO PROCESSING 1 An experimental comparison of audio tempo Uhle Abstract-- We report on the tempo induction contest organ- ised during the International the basic tempo (as a scalar, in beats per minute) from musical audio signals. To our knowledge

  19. DRAFT FOR IEEE TRANS. ON SPEECH AND AUDIO PROCESSING 1 An experimental comparison of audio tempo

    E-print Network

    Dixon, Simon

    DRAFT FOR IEEE TRANS. ON SPEECH AND AUDIO PROCESSING 1 An experimental comparison of audio tempo Uhle, Pedro Cano Abstract-- We report on the tempo induction contest organ- ised during in the task of inducing the basic tempo (as a scalar, in beats per minute) from musical audio signals. To our

  20. Three-Dimensional Audio Client Library

    NASA Technical Reports Server (NTRS)

    Rizzi, Stephen A.

    2005-01-01

    The Three-Dimensional Audio Client Library (3DAudio library) is a group of software routines written to facilitate development of both stand-alone (audio only) and immersive virtual-reality application programs that utilize three-dimensional audio displays. The library is intended to enable the development of three-dimensional audio client application programs by use of a code base common to multiple audio server computers. The 3DAudio library calls vendor-specific audio client libraries and currently supports the AuSIM Gold-Server and Lake Huron audio servers. 3DAudio library routines contain common functions for (1) initiation and termination of a client/audio server session, (2) configuration-file input, (3) positioning functions, (4) coordinate transformations, (5) audio transport functions, (6) rendering functions, (7) debugging functions, and (8) event-list-sequencing functions. The 3DAudio software is written in the C++ programming language and currently operates under the Linux, IRIX, and Windows operating systems.

  1. Collusion-resistant audio fingerprinting system in the modulated complex lapped transform domain.

    PubMed

    Garcia-Hernandez, Jose Juan; Feregrino-Uribe, Claudia; Cumplido, Rene

    2013-01-01

    Collusion-resistant fingerprinting paradigm seems to be a practical solution to the piracy problem as it allows media owners to detect any unauthorized copy and trace it back to the dishonest users. Despite the billionaire losses in the music industry, most of the collusion-resistant fingerprinting systems are devoted to digital images and very few to audio signals. In this paper, state-of-the-art collusion-resistant fingerprinting ideas are extended to audio signals and the corresponding parameters and operation conditions are proposed. Moreover, in order to carry out fingerprint detection using just a fraction of the pirate audio clip, block-based embedding and its corresponding detector is proposed. Extensive simulations show the robustness of the proposed system against average collusion attack. Moreover, by using an efficient Fast Fourier Transform core and standard computer machines it is shown that the proposed system is suitable for real-world scenarios. PMID:23762455

  2. Collusion-Resistant Audio Fingerprinting System in the Modulated Complex Lapped Transform Domain

    PubMed Central

    Garcia-Hernandez, Jose Juan; Feregrino-Uribe, Claudia; Cumplido, Rene

    2013-01-01

    Collusion-resistant fingerprinting paradigm seems to be a practical solution to the piracy problem as it allows media owners to detect any unauthorized copy and trace it back to the dishonest users. Despite the billionaire losses in the music industry, most of the collusion-resistant fingerprinting systems are devoted to digital images and very few to audio signals. In this paper, state-of-the-art collusion-resistant fingerprinting ideas are extended to audio signals and the corresponding parameters and operation conditions are proposed. Moreover, in order to carry out fingerprint detection using just a fraction of the pirate audio clip, block-based embedding and its corresponding detector is proposed. Extensive simulations show the robustness of the proposed system against average collusion attack. Moreover, by using an efficient Fast Fourier Transform core and standard computer machines it is shown that the proposed system is suitable for real-world scenarios. PMID:23762455

  3. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ...radio station must simulcast its analog audio programming on one of its digital audio programming streams. The DAB audio programming stream that is provided...must be at least comparable in sound quality to the analog programming service...

  4. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ...radio station must simulcast its analog audio programming on one of its digital audio programming streams. The DAB audio programming stream that is provided...must be at least comparable in sound quality to the analog programming service...

  5. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ...radio station must simulcast its analog audio programming on one of its digital audio programming streams. The DAB audio programming stream that is provided...must be at least comparable in sound quality to the analog programming service...

  6. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ...radio station must simulcast its analog audio programming on one of its digital audio programming streams. The DAB audio programming stream that is provided...must be at least comparable in sound quality to the analog programming service...

  7. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ...radio station must simulcast its analog audio programming on one of its digital audio programming streams. The DAB audio programming stream that is provided...must be at least comparable in sound quality to the analog programming service...

  8. VISUAL AND AUDIO PRESENTATION IN MACHINE PROGRAMED INSTRUCTION. FINAL REPORT.

    ERIC Educational Resources Information Center

    ALLEN, WILLIAM H.

    THIS STUDY WAS PART OF A LARGER RESEARCH PROGRAM AIMED TOWARD DEVELOPMENT OF PARADIGMS OF MESSAGE DESIGN. OBJECTIVES OF THREE PARALLEL EXPERIMENTS WERE TO EVALUATE INTERACTIONS OF PRESENTATION MODE, PROGRAM TYPE, AND CONTENT AS THEY AFFECT LEARNER CHARACTERISTICS. EACH EXPERIMENT USED 18 TREATMENTS IN A FACTORIAL DESIGN WITH RANDOMLY SELECTED…

  9. Improving Audio Quality in Distance Learning Applications.

    ERIC Educational Resources Information Center

    Richardson, Craig H.

    This paper discusses common causes of problems encountered with audio systems in distance learning networks and offers practical suggestions for correcting the problems. Problems and discussions are divided into nine categories: (1) acoustics, including reverberant classrooms leading to distorted or garbled voices, as well as one-dimensional audio

  10. Digital Advances in Contemporary Audio Production.

    ERIC Educational Resources Information Center

    Shields, Steven O.

    Noting that a revolution in sonic high fidelity occurred during the 1980s as digital-based audio production methods began to replace traditional analog modes, this paper offers both an overview of digital audio theory and descriptions of some of the related digital production technologies that have begun to emerge from the mating of the computer…

  11. Technical Advances in Digital Audio Radio Broadcasting

    E-print Network

    Juang, Biing-Hwang "Fred"

    such as Eureka-147, DRM, and Worldspace, developed under different re- quirements. Keywords--Audio coding of electric current over con- ducting wires. Marconi's transmission took the form of Morse code, which in television distribution. This is no less true for audio broadcasting which has taken a unique development

  12. Digital Audio Sampling for Film and Video.

    ERIC Educational Resources Information Center

    Stanton, Michael J.

    Digital audio sampling is explained, and some of its implications in digital sound applications are discussed. Digital sound equipment is rapidly replacing analog recording devices as the state-of-the-art in audio technology. The philosophy of digital recording involves doing away with the continuously variable analog waveforms and turning the…

  13. Digital Audio: A Sound Design Element.

    ERIC Educational Resources Information Center

    Barron, Ann; Varnadoe, Susan

    1992-01-01

    Discussion of incorporating audio into videodiscs for multimedia educational applications highlights a project developed for the Navy that used digital audio in an interactive video delivery system (IVDS) for training sonar operators. Storage constraints with videodiscs are explained, design requirements for the IVDS are described, and production…

  14. AUDIO FORENSICS FROM ACOUSTIC REVERBERATION Hafiz Malik

    E-print Network

    Bucci, David J.

    AUDIO FORENSICS FROM ACOUSTIC REVERBERATION Hafiz Malik Department of Electrical and Computer be used in a forensic and ballistic setting. Index Terms-- Audio Forensics 1. INTRODUCTION The past few years have seen significant advances in image forensics [1]. At the same time, techniques

  15. Automatic Audio and Lyrics Alignment DIPLOMARBEIT

    E-print Network

    Widmer, Gerhard

    Automatic Audio and Lyrics Alignment DIPLOMARBEIT zur Erlangung des akademischen Grades Diplom for the song currently played. One kind of information is the lyrics; that is what my diploma thesis deals with. The goal is to provide a pro- gram that is able to automatically align the lyrics to the audio signal

  16. Statistical Measurement Approach for On-line Audio Quality Assessment

    E-print Network

    Wang, Yongge

    Statistical Measurement Approach for On-line Audio Quality Assessment Lopamudra Roychoudhuri, Ehab, Depaul University, 243 S. Wabash Ave., Chicago, IL-60604, U.S.A Abstract. The quality of audio in IP on audio compression techniques. On-line audio quality assessment is impor- tant to provide real

  17. INTELLIGENT SYSTEMS FOR MIXING MULTICHANNEL AUDIO Joshua D. Reiss

    E-print Network

    Reiss, Josh

    in audio production quality. This issue has been addressed in the context of audio source separationINTELLIGENT SYSTEMS FOR MIXING MULTICHANNEL AUDIO Joshua D. Reiss Centre for Digital Music, Queen field of multichannel audio signal processing where the inter-channel relationships are exploited

  18. Audio Transmission over the Internet: Experiments and Observations*

    E-print Network

    Wang, Yongge

    to maintain the quality of the audio in a multimedia transmission. In the absence of guarantee of quality tested for subjective audio quality analysis and to observe the effects of network condition on different differently in terms of audio quality. Thus the quality of an audio transmission depends not only

  19. A new audio device driver abstraction Dickon Reed

    E-print Network

    Hand, Steven

    services. High quality audio has, compared to other media types, relatively low band­ width, but is veryA new audio device driver abstraction Dickon Reed Computer Laboratory, University of Cambridge­ ence ``Networks and Operating Systems Support for distributed audio and video'', 1998 Traditional audio

  20. USING 3D AUDIO GUIDANCE TO LOCATE INDOOR STATIC OBJECTS

    E-print Network

    Tsigas, Philippas

    ", etc. The 3D audio effect is very subtle and thus requires high quality audio with a carefully chosenUSING 3D AUDIO GUIDANCE TO LOCATE INDOOR STATIC OBJECTS Samuel Sandberg Email: kv00ssg University of Technology Is 3D audio an interesting technology for displaying navigational information

  1. AME CURRICULUM GUIDE Fall 2015 AUDIO AND MUSIC ENGINEERING

    E-print Network

    Cantlon, Jessica F.

    choose to earn a Master of Science degree in Electrical Engineering with as little as one additional year manufacturing, audio software development, music and audio recording and production, core audio and signal processing technologies and component manufacturing, musical instruments and audio equipment manufacturing

  2. High-Fidelity Piezoelectric Audio Device

    NASA Technical Reports Server (NTRS)

    Woodward, Stanley E.; Fox, Robert L.; Bryant, Robert G.

    2003-01-01

    ModalMax is a very innovative means of harnessing the vibration of a piezoelectric actuator to produce an energy efficient low-profile device with high-bandwidth high-fidelity audio response. The piezoelectric audio device outperforms many commercially available speakers made using speaker cones. The piezoelectric device weighs substantially less (4 g) than the speaker cones which use magnets (10 g). ModalMax devices have extreme fabrication simplicity. The entire audio device is fabricated by lamination. The simplicity of the design lends itself to lower cost. The piezoelectric audio device can be used without its acoustic chambers and thereby resulting in a very low thickness of 0.023 in. (0.58 mm). The piezoelectric audio device can be completely encapsulated, which makes it very attractive for use in wet environments. Encapsulation does not significantly alter the audio response. Its small size (see Figure 1) is applicable to many consumer electronic products, such as pagers, portable radios, headphones, laptop computers, computer monitors, toys, and electronic games. The audio device can also be used in automobile or aircraft sound systems.

  3. Ye Wang and Miikka Vilermo/Exploiting Excess Masking for Audio Compression International Conference on High Quality Audio Coding 1

    E-print Network

    Wang, Ye

    Conference on High Quality Audio Coding 1 EXPLOITING EXCESS MASKING FOR AUDIO COMPRESSION YE WANG AND MIIKKAYe Wang and Miikka Vilermo/Exploiting Excess Masking for Audio Compression AES 17 th International VILERMO Nokia Research Center Speech and Audio Systems Lab. Tampere, Finland ye

  4. 1014 IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 13, NO. 5, SEPTEMBER 2005 Robust Low-Delay Audio Coding

    E-print Network

    Kova?evi?, Jelena

    for high-quality, low-delay audio communication that is robust to losses in packetized transmission audio quality--than simple alternative coders with the same low delay. Index Terms--Audio coding, error with high quality for both the video and the audio, musicians playing together remotely, wireless speakers

  5. Content-based access to spoken audio 

    E-print Network

    Koumpis, Konstantinos; Renals, Steve

    2005-01-01

    The amount of archived audio material in digital form is increasing rapidly, as advantage is taken of the growth in available storage and processing power. Computational resources are becoming less of a bottleneck to ...

  6. Audio information access from meeting rooms. 

    E-print Network

    Renals, Steve; Ellis, Dan

    2003-01-01

    We investigate approaches to accessing information from the streams of audio data that result from multi-channel recordings of meetings. The methods investigated use word-level transcriptions, and information derived ...

  7. Web Audio/Video Streaming Tool

    NASA Technical Reports Server (NTRS)

    Guruvadoo, Eranna K.

    2003-01-01

    In order to promote NASA-wide educational outreach program to educate and inform the public of space exploration, NASA, at Kennedy Space Center, is seeking efficient ways to add more contents to the web by streaming audio/video files. This project proposes a high level overview of a framework for the creation, management, and scheduling of audio/video assets over the web. To support short-term goals, the prototype of a web-based tool is designed and demonstrated to automate the process of streaming audio/video files. The tool provides web-enabled users interfaces to manage video assets, create publishable schedules of video assets for streaming, and schedule the streaming events. These operations are performed on user-defined and system-derived metadata of audio/video assets stored in a relational database while the assets reside on separate repository. The prototype tool is designed using ColdFusion 5.0.

  8. TRANSCRIBING AUDIO-VIDEO ARCHIVES Claude Barras

    E-print Network

    TRANSCRIBING AUDIO-VIDEO ARCHIVES Claude Barras½ , Alexandre Allauzen½ ¾ , Lori Lamel½ , Jean-sur-Marne cedex, France barras,allauzen,lamel,gauvain @limsi.fr ABSTRACT This paper addresses the automatic

  9. 36 CFR 1002.12 - Audio disturbances.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ...prohibited: (1) Operating motorized equipment or machinery such as an electric generating plant, motor vehicle, motorized toy, or an audio device, such as a radio, television set, tape deck or musical instrument, in a manner that exceeds...

  10. 36 CFR 2.12 - Audio disturbances.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ...prohibited: (1) Operating motorized equipment or machinery such as an electric generating plant, motor vehicle, motorized toy, or an audio device, such as a radio, television set, tape deck or musical instrument, in a manner: (i) That...

  11. Huffman coding in advanced audio coding standard

    NASA Astrophysics Data System (ADS)

    Brzuchalski, Grzegorz

    2012-05-01

    This article presents several hardware architectures of Advanced Audio Coding (AAC) Huffman noiseless encoder, its optimisations and working implementation. Much attention has been paid to optimise the demand of hardware resources especially memory size. The aim of design was to get as short binary stream as possible in this standard. The Huffman encoder with whole audio-video system has been implemented in FPGA devices.

  12. An objective model for audio-visual quality

    NASA Astrophysics Data System (ADS)

    Martinez, Helard Becerra; Farias, Mylène C. Q.

    2014-01-01

    In this paper, we describe three psychophysical experiments with the goal of understanding the influence of audio and video components on the overall perceived audio- visual quality. In Experiment I, subjects evaluated the quality of videos (without any audio) compressed at different video bitrates. In Experiment II, subjects evaluated the quality of audio (without any video) compressed at different audio bitrates. In Experiment III, subjects evaluated the quality of videos (audio- visual signals), which had their audio and video components compressed at different bitrates. The results of these experiments show that compressing only the video have a higher impact on the overall perceived quality than compressing only the audio. Another important goal of this paper is to propose an objective model for the audio-visual quality. Using the gathered data from Experiments I, II, and III, we are able to obtain two models with reasonably good correlations with the overall perceived quality.

  13. MACHINE CONTEST MACHINE VOLUME WORKSHEET

    E-print Network

    Giles, C. Lee

    MACHINE CONTEST® MACHINE VOLUME WORKSHEET Note: Each square is 1' x 1' (0.3 m x 0.3 m) (Official the footprint of your machine NOTE: If any part of the machine falls within a square, you must count the entire: _______________ (choose one: ft2 or m2 ) STEP 3: Measure the height of your machine. NOTE: The maximum height is 8' (2.4 m

  14. Could Audio-Described Films Benefit from Audio Introductions? An Audience Response Study

    ERIC Educational Resources Information Center

    Romero-Fresco, Pablo; Fryer, Louise

    2013-01-01

    Introduction: Time constraints limit the quantity and type of information conveyed in audio description (AD) for films, in particular the cinematic aspects. Inspired by introductory notes for theatre AD, this study developed audio introductions (AIs) for "Slumdog Millionaire" and "Man on Wire." Each AI comprised 10 minutes of…

  15. Long Range Audio and Audio-Visual Event Detection Using a Laser Doppler Vibrometer

    E-print Network

    Zhu, Zhigang

    , is a long-range, non-contact acoustic measurement device to detect the speed of the target's vibration basedLong Range Audio and Audio-Visual Event Detection Using a Laser Doppler Vibrometer Tao Wangab from a large distance. Setting up a long range microphone array and performing geo-calibration of both

  16. High quality scalable audio codec

    NASA Astrophysics Data System (ADS)

    Kim, Miyoung; Oh, Eunmi; Kim, JungHoe

    2007-09-01

    The MPEG-4 BSAC (Bit Sliced Arithmetic Coding) is a fine-grain scalable codec with layered structure which consists of a single base-layer and several enhancement layers. The scalable functionality allows us to decode the subsets of a full bitstream and to deliver audio contents adaptively under conditions of heterogeneous network and devices, and user interaction. This bitrate scalability can be provided at the cost of high frequency components. It means that the decoded output of BSAC sounds muffled as the transmitted layers become less and less due to deprived conditions of network and devices. The goal of the proposed technology is to compensate the missing high frequency components, while maintaining the fine grain scalability of BSAC. This paper describes the integration of SBR (Spectral Bandwidth Replication) tool to existing MPEG-4 BSAC. Listening test results show that the sound quality of BSAC is improved when the full bitstream is truncated for lower bitrates, and this quality is comparable to that of BSAC using SBR tool without truncation at the same bitrate.

  17. Digital Multicasting of Multiple Audio Streams

    NASA Technical Reports Server (NTRS)

    Macha, Mitchell; Bullock, John

    2007-01-01

    The Mission Control Center Voice Over Internet Protocol (MCC VOIP) system (see figure) comprises hardware and software that effect simultaneous, nearly real-time transmission of as many as 14 different audio streams to authorized listeners via the MCC intranet and/or the Internet. The original version of the MCC VOIP system was conceived to enable flight-support personnel located in offices outside a spacecraft mission control center to monitor audio loops within the mission control center. Different versions of the MCC VOIP system could be used for a variety of public and commercial purposes - for example, to enable members of the general public to monitor one or more NASA audio streams through their home computers, to enable air-traffic supervisors to monitor communication between airline pilots and air-traffic controllers in training, and to monitor conferences among brokers in a stock exchange. At the transmitting end, the audio-distribution process begins with feeding the audio signals to analog-to-digital converters. The resulting digital streams are sent through the MCC intranet, using a user datagram protocol (UDP), to a server that converts them to encrypted data packets. The encrypted data packets are then routed to the personal computers of authorized users by use of multicasting techniques. The total data-processing load on the portion of the system upstream of and including the encryption server is the total load imposed by all of the audio streams being encoded, regardless of the number of the listeners or the number of streams being monitored concurrently by the listeners. The personal computer of a user authorized to listen is equipped with special- purpose MCC audio-player software. When the user launches the program, the user is prompted to provide identification and a password. In one of two access- control provisions, the program is hard-coded to validate the user s identity and password against a list maintained on a domain-controller computer at the MCC. In the other access-control provision, the program verifies that the user is authorized to have access to the audio streams. Once both access-control checks are completed, the audio software presents a graphical display that includes audiostream-selection buttons and volume-control sliders. The user can select all or any subset of the available audio streams and can adjust the volume of each stream independently of that of the other streams. The audio-player program spawns a "read" process for the selected stream(s). The spawned process sends, to the router(s), a "multicast-join" request for the selected streams. The router(s) responds to the request by sending the encrypted multicast packets to the spawned process. The spawned process receives the encrypted multicast packets and sends a decryption packet to audio-driver software. As the volume or muting features are changed by the user, interrupts are sent to the spawned process to change the corresponding attributes sent to the audio-driver software. The total latency of this system - that is, the total time from the origination of the audio signals to generation of sound at a listener s computer - lies between four and six seconds.

  18. Acoustic Cues to Beat Induction: A Machine Learning Perspective

    E-print Network

    Widmer, Gerhard

    music pieces. We consider many different features computed on consecutive short portions of acoustic their ability to provide reliable cues to the localization of beats is based on a machine learning methodology with a large corpus of beat-annotated music pieces, in audio format, covering distinctive music categories

  19. Spatial domain entertainment audio decompression/compression

    NASA Astrophysics Data System (ADS)

    Chan, Y. K.; Tam, Ka Him K.

    2014-02-01

    The ARM7 NEON processor with 128bit SIMD hardware accelerator requires a peak performance of 13.99 Mega Cycles per Second for MP3 stereo entertainment quality decoding. For similar compression bit rate, OGG and AAC is preferred over MP3. The Patent Cooperation Treaty Application dated 28/August/2012 describes an audio decompression scheme producing a sequence of interleaving "min to Max" and "Max to min" rising and falling segments. The number of interior audio samples bound by "min to Max" or "Max to min" can be {0|1|…|N} audio samples. The magnitudes of samples, including the bounding min and Max, are distributed as normalized constants within the 0 and 1 of the bounding magnitudes. The decompressed audio is then a "sequence of static segments" on a frame by frame basis. Some of these frames needed to be post processed to elevate high frequency. The post processing is compression efficiency neutral and the additional decoding complexity is only a small fraction of the overall decoding complexity without the need of extra hardware. Compression efficiency can be speculated as very high as source audio had been decimated and converted to a set of data with only "segment length and corresponding segment magnitude" attributes. The PCT describes how these two attributes are efficiently coded by the PCT innovative coding scheme. The PCT decoding efficiency is obviously very high and decoding latency is basically zero. Both hardware requirement and run time is at least an order of magnitude better than MP3 variants. The side benefit is ultra low power consumption on mobile device. The acid test on how such a simplistic waveform representation can indeed reproduce authentic decompressed quality is benchmarked versus OGG(aoTuv Beta 6.03) by three pair of stereo audio frames and one broadcast like voice audio frame with each frame consisting 2,028 samples at 44,100KHz sampling frequency.

  20. Audio stream classification for multimedia database search

    NASA Astrophysics Data System (ADS)

    Artese, M.; Bianco, S.; Gagliardi, I.; Gasparini, F.

    2013-03-01

    Search and retrieval of huge archives of Multimedia data is a challenging task. A classification step is often used to reduce the number of entries on which to perform the subsequent search. In particular, when new entries of the database are continuously added, a fast classification based on simple threshold evaluation is desirable. In this work we present a CART-based (Classification And Regression Tree [1]) classification framework for audio streams belonging to multimedia databases. The database considered is the Archive of Ethnography and Social History (AESS) [2], which is mainly composed of popular songs and other audio records describing the popular traditions handed down generation by generation, such as traditional fairs, and customs. The peculiarities of this database are that it is continuously updated; the audio recordings are acquired in unconstrained environment; and for the non-expert human user is difficult to create the ground truth labels. In our experiments, half of all the available audio files have been randomly extracted and used as training set. The remaining ones have been used as test set. The classifier has been trained to distinguish among three different classes: speech, music, and song. All the audio files in the dataset have been previously manually labeled into the three classes above defined by domain experts.

  1. Interaction with Machine Improvisation

    NASA Astrophysics Data System (ADS)

    Assayag, Gerard; Bloch, George; Cont, Arshia; Dubnov, Shlomo

    We describe two multi-agent architectures for an improvisation oriented musician-machine interaction systems that learn in real time from human performers. The improvisation kernel is based on sequence modeling and statistical learning. We present two frameworks of interaction with this kernel. In the first, the stylistic interaction is guided by a human operator in front of an interactive computer environment. In the second framework, the stylistic interaction is delegated to machine intelligence and therefore, knowledge propagation and decision are taken care of by the computer alone. The first framework involves a hybrid architecture using two popular composition/performance environments, Max and OpenMusic, that are put to work and communicate together, each one handling the process at a different time/memory scale. The second framework shares the same representational schemes with the first but uses an Active Learning architecture based on collaborative, competitive and memory-based learning to handle stylistic interactions. Both systems are capable of processing real-time audio/video as well as MIDI. After discussing the general cognitive background of improvisation practices, the statistical modelling tools and the concurrent agent architecture are presented. Then, an Active Learning scheme is described and considered in terms of using different improvisation regimes for improvisation planning. Finally, we provide more details about the different system implementations and describe several performances with the system.

  2. The Effect of Audio and Visual Aids on Task Performance in Distributed Collaborative Virtual Environments

    NASA Astrophysics Data System (ADS)

    Ullah, Sehat; Richard, Paul; Otman, Samir; Mallem, Malik

    2009-03-01

    Collaborative virtual environments (CVE) has recently gained the attention of many researchers due to its numerous potential application domains. Cooperative virtual environments, where users simultaneously manipulate objects, is one of the subfields of CVEs. In this paper we present a framework that enables two users to cooperatively manipulate objects in virtual environment, while setting on two separate machines connected through local network. In addition the article presents the use of sensory feedback (audio and visual) and investigates their effects on the cooperation and user's performance. Six volunteers subject had to cooperatively perform a peg-in-hole task. Results revealed that visual and auditory aid increase users' performance. However majority of the users preferred visual feedback to audio. We hope this framework will greatly help in the development of CAD systems that allow the designers to collaboratively design while being distant. Similarly other application domains may be cooperative assembly, surgical training and rehabilitation systems.

  3. Podscanning : audio microcontent and synchronous communication for mobile devices

    E-print Network

    Wheeler, Patrick Sean

    2010-01-01

    Over the past decade, computationally powerful audio communication devices have become commonplace. Mobile devices have high storage capacity for digital audio, and smartphones or networked PDAs can be used to stream ...

  4. Audio compression with non-uniform modulated complex lapped transform 

    E-print Network

    Scheuble, Anne-Sophie Maud

    2000-01-01

    In the past 10 years, advances in audio compression have been tremendous. Lapped Orthogonal Transforms (LOT) and psychoacoustic properties have significantly improved performance of transform audio coders. As a result, standards like MP3...

  5. Development and Testing of Achievement from Multiple Modes of Mathematical Representation: Audio, Audio-Visual, and Kinesthetic 

    E-print Network

    Ozel, Serkan

    2012-02-14

    by allocated session time (10-, 20-, and 30-minute). The online manipulative tool (OMT), which was designed to scaffold learning in operations with rational numbers, allowed students to use the following three components in any order: (a) audio, (b) audio...

  6. 47 CFR 10.520 - Common audio attention signal.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 47 Telecommunication 1 2013-10-01 2013-10-01 false Common audio attention signal. 10.520 Section... Equipment Requirements § 10.520 Common audio attention signal. A Participating CMS Provider and equipment manufacturers may only market devices for public use under part 10 that include an audio attention signal...

  7. 47 CFR 10.520 - Common audio attention signal.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 47 Telecommunication 1 2012-10-01 2012-10-01 false Common audio attention signal. 10.520 Section... Equipment Requirements § 10.520 Common audio attention signal. A Participating CMS Provider and equipment manufacturers may only market devices for public use under part 10 that include an audio attention signal...

  8. 47 CFR 10.520 - Common audio attention signal.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 47 Telecommunication 1 2011-10-01 2011-10-01 false Common audio attention signal. 10.520 Section... Equipment Requirements § 10.520 Common audio attention signal. A Participating CMS Provider and equipment manufacturers may only market devices for public use under part 10 that include an audio attention signal...

  9. 47 CFR 10.520 - Common audio attention signal.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 47 Telecommunication 1 2014-10-01 2014-10-01 false Common audio attention signal. 10.520 Section... Equipment Requirements § 10.520 Common audio attention signal. A Participating CMS Provider and equipment manufacturers may only market devices for public use under part 10 that include an audio attention signal...

  10. Object-Coding for Resolution-Free Musical Audio

    E-print Network

    Plumbley, Mark

    quality" audio currently in use (as samples, digital recordings and CDs) which need to be fittedObject-Coding for Resolution-Free Musical Audio Stephen J. Welburn1, Mark D. Plumbley1 and Emmanuel be addressed to Steve Welburn (stephen.welburn@elec.qmul.ac.uk) ABSTRACT Object-based coding of audio

  11. Queen Mary's `Media & Arts Technology Studios' Audio System Design

    E-print Network

    Reiss, Josh

    and are recorded by engineers in the Control Room. This will require high quality multi- channel audio recordingQueen Mary's `Media & Arts Technology Studios' Audio System Design Martin J. Morrell1 , Christopher discusses our design philosophy for our given brief to create a world-class audio recording

  12. Audio signal representations for indexing in the transform domain

    E-print Network

    Paris 7 - Denis Diderot, Université

    little) loss in the perceived quality of the decoded signal. The basic principle of an audio coder1 Audio signal representations for indexing in the transform domain Emmanuel Ravelli, Student--Indexing audio signals directly in the transform domain can potentially save a significant amount of computation

  13. THE AUDIO DEGRADATION TOOLBOX AND ITS APPLICATION TO ROBUSTNESS EVALUATION

    E-print Network

    Mauch, Matthias

    degradation. For example, audio is degraded by low-quality microphones, noisy recording environments, MP3 methods are robust against degradations in audio quality. For example, the robustness of automatic speech informatics only a few studies consider insufficient or variable audio quality, exceptions including two

  14. An MDCT Hardware Accelerator for MP3 Audio Xingdong Dai

    E-print Network

    Wagh, Meghanad

    transforms is im- portant to the realization of high quality audio compression, especially when most MP3An MDCT Hardware Accelerator for MP3 Audio Xingdong Dai LSI Corporation 1110 American Parkway NE mdw0@lehigh.edu Abstract-- With the increasing popularity of MP3 audio, there is a need to develop

  15. PRELIMINARY GUIDELINES FOR SUBJECTIVE EVALUATION OF AUDIO SOURCE SEPARATION ALGORITHMS

    E-print Network

    Plumbley, Mark

    .lastname@elec.qmul.ac.uk ABSTRACT Evaluating audio source separation algorithms means rat- ing the quality or intelligibility to evaluate the basic audio quality of separated sources and provide an example of their application using audio quality in many cases [19], they do not model auditory phenomena of loudness weighting

  16. Real-time Audio Quality Evaluation for Adaptive Multimedia Protocols

    E-print Network

    Wang, Yongge

    Real-time Audio Quality Evaluation for Adaptive Multimedia Protocols Lopamudra RoychoudhuriPaul University, 243 S. Wabash Ave., Chicago, IL-60604, U.S.A Abstract. The quality of audio in IP telephony compression technique. In this paper, we propose a novel statistical-based on-line audio quality assessment

  17. Optimal Audio Transmission over Error-Prone Wireless Links

    E-print Network

    Yousefi'zadeh, Homayoun

    for transmitting high quality audio sequences over error-prone wireless links. Our framework introduces apparatus due to an ever increasing de- mand. In that context, delivering high quality audio over error1 Optimal Audio Transmission over Error-Prone Wireless Links Ala' Khalifeh, Student Member IEEE

  18. MULTICHANNEL AUDIO CODING USING SINUSOIDAL MODELLING AND COMPRESSED SENSING

    E-print Network

    Tsakalides, Panagiotis

    the sinusoidal part of an audio signal pro- vides sufficient quality. It is shown here that, except from oneMULTICHANNEL AUDIO CODING USING SINUSOIDAL MODELLING AND COMPRESSED SENSING Anthony Griffin, Toni the potential of applying com- pressed sensing (CS) to multichannel audio coding. In this context, we consider

  19. Virtualized Audio as a Distributed Interactive Application Peter A. Dinda

    E-print Network

    Dinda, Peter A.

    - alized audio would require only that a user have a collec- tion of high quality microphones and a high quality set of headphones. However, the virtualized audio model is not limited to this equipment, and weVirtualized Audio as a Distributed Interactive Application Peter A. Dinda Department of Computer

  20. Audio signal representations for indexing in the transform domain

    E-print Network

    Richard, Gaël

    a PCM audio signal with no (or little) loss in the perceived quality of the decoded signal. The basic the size of a PCM audio signal more than 6 times while guaranteeing a near-transparent quality1 Audio signal representations for indexing in the transform domain Emmanuel Ravelli, Ga¨el Richard

  1. Transparent and Robust Audio Data Hiding in Subband Domain

    E-print Network

    Li, Xin

    range of attacks while providing transparent audio quality and abundant embedding capacity(>20bps). 1 on audio quality but could seriously affect the extraction of embedded data. They have shown to be the mostTransparent and Robust Audio Data Hiding in Subband Domain Xin Li1* and Hong Heather Yu2 1 Dept

  2. Semiconductors Audio circuits using the NE5532/3/4

    E-print Network

    Lanterman, Aaron

    for application in high quality and professional audio equipment, instrumentation and control circuits for use in high quality and professional audio equipment which requires low noise and low distortionPhilips Semiconductors AN142 Audio circuits using the NE5532/3/4 author 1984 Oct INTEGRATED

  3. Adaptive Wavelet Quantization Index Modulation Technique for Audio Watermarking

    E-print Network

    Chang, Pao-Chi

    system can deliver perceptual transparent audio quality, and it is robust against various signalAdaptive Wavelet Quantization Index Modulation Technique for Audio Watermarking Jong-Tzy Wang 1. In this paper, we propose a robust audio watermarking technique which adopts the wavelet QIM method

  4. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 47 Telecommunication 4 2012-10-01 2012-10-01 false Digital audio broadcasting service requirements. 73.403 Section 73.403 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST RADIO SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast...

  5. 47 CFR 10.520 - Common audio attention signal.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 47 Telecommunication 1 2010-10-01 2010-10-01 false Common audio attention signal. 10.520 Section... Equipment Requirements § 10.520 Common audio attention signal. A Participating CMS Provider and equipment manufacturers may only market devices for public use under part 10 that include an audio attention signal...

  6. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 47 Telecommunication 4 2010-10-01 2010-10-01 false Digital audio broadcasting service requirements... SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast radio stations using IBOC must transmit at least one over-the-air digital...

  7. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 47 Telecommunication 4 2013-10-01 2013-10-01 false Digital audio broadcasting service requirements... SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast radio stations using IBOC must transmit at least one over-the-air digital...

  8. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 47 Telecommunication 4 2011-10-01 2011-10-01 false Digital audio broadcasting service requirements... SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast radio stations using IBOC must transmit at least one over-the-air digital...

  9. 47 CFR 73.403 - Digital audio broadcasting service requirements.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 47 Telecommunication 4 2014-10-01 2014-10-01 false Digital audio broadcasting service requirements... SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast radio stations using IBOC must transmit at least one over-the-air digital...

  10. Nonlinear dynamic macromodeling techniques for audio systems

    NASA Astrophysics Data System (ADS)

    Ogrodzki, Jan; Bie?kowski, Piotr

    2015-09-01

    This paper develops a modelling method and a models identification technique for the nonlinear dynamic audio systems. Identification is performed by means of a behavioral approach based on a polynomial approximation. This approach makes use of Discrete Fourier Transform and Harmonic Balance Method. A model of an audio system is first created and identified and then it is simulated in real time using an algorithm of low computational complexity. The algorithm consists in real time emulation of the system response rather than in simulation of the system itself. The proposed software is written in Python language using object oriented programming techniques. The code is optimized for a multithreads environment.

  11. Enhancing Navigation Skills through Audio Gaming

    PubMed Central

    Sánchez, Jaime; Sáenz, Mauricio; Pascual-Leone, Alvaro; Merabet, Lotfi

    2014-01-01

    We present the design, development and initial cognitive evaluation of an Audio-based Environment Simulator (AbES). This software allows a blind user to navigate through a virtual representation of a real space for the purposes of training orientation and mobility skills. Our findings indicate that users feel satisfied and self-confident when interacting with the audio-based interface, and the embedded sounds allow them to correctly orient themselves and navigate within the virtual world. Furthermore, users are able to transfer spatial information acquired through virtual interactions into real world navigation and problem solving tasks. PMID:25505796

  12. Text-to-Speech and Reading While Listening: Reading Support for Individuals with Severe Traumatic Brain Injury

    ERIC Educational Resources Information Center

    Harvey, Judy

    2013-01-01

    Individuals with severe traumatic brain injury (TBI) often have reading challenges. They maintain or reestablish basic decoding and word recognition skills following injury, but problems with reading comprehension often persist. Practitioners have the potential to accommodate struggling readers by changing the presentational mode of text in a…

  13. Building Digital Audio Preservation Infrastructure and Workflows

    ERIC Educational Resources Information Center

    Young, Anjanette; Olivieri, Blynne; Eckler, Karl; Gerontakos, Theodore

    2010-01-01

    In 2009 the University of Washington (UW) Libraries special collections received funding for the digital preservation of its audio indigenous language holdings. The university libraries, where the authors work in various capacities, had begun digitizing image and text collections in 1997. Because of this, at the onset of the project, workflows (a…

  14. Audio Filters What you will do

    E-print Network

    Anderson, Betty Lise

    how much current flows · Protect components #12;And this? Capacitor Stops DC, passes AC The lowerAudio Filters #12;What you will do Build a 3-band equalizer Low pass filter High pass filter Band pass filter Connect to a music source (mp3 player) Adjust the strength of low, high, and middle

  15. Agency Video, Audio and Imagery Library

    NASA Technical Reports Server (NTRS)

    Grubbs, Rodney

    2015-01-01

    The purpose of this presentation was to inform the ISS International Partners of the new NASA Agency Video, Audio and Imagery Library (AVAIL) website. AVAIL is a new resource for the public to search for and download NASA-related imagery, and is not intended to replace the current process by which the International Partners receive their Space Station imagery products.

  16. Audio-Visual Speech Perception Is Special

    ERIC Educational Resources Information Center

    Tuomainen, J.; Andersen, T.S.; Tiippana, K.; Sams, M.

    2005-01-01

    In face-to-face conversation speech is perceived by ear and eye. We studied the prerequisites of audio-visual speech perception by using perceptually ambiguous sine wave replicas of natural speech as auditory stimuli. When the subjects were not aware that the auditory stimuli were speech, they showed only negligible integration of auditory and…

  17. Spanish for Agricultural Purposes: The Audio Program.

    ERIC Educational Resources Information Center

    Mainous, Bruce H.; And Others

    The manual is meant to accompany and supplement the basic manual and to serve as support to the audio component of "Spanish for Agricultural Purposes," a one-semester course for North American agriculture specialists preparing to work in Latin America, consists of exercises to supplement readings presented in the course's basic manual and to…

  18. Neuroprocessing: Image Compres-sion & Audio Processing

    E-print Network

    Dony, Bob

    Components: Hebbian Learning NN w11 y1 ym xn x2 x1 w12 wm1 wmn w1n w2m Neuroprocessing: Applications in Image Compression and Audio Processing ­ p.13/42 #12;APEX Network x1 x2 x3 xN y1 y2 y3 ym w c y Neuroprocessing

  19. 50 CFR 27.72 - Audio equipment.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 50 Wildlife and Fisheries 9 2014-10-01 2014-10-01 false Audio equipment. 27.72 Section 27.72 Wildlife and Fisheries UNITED STATES FISH AND WILDLIFE SERVICE, DEPARTMENT OF THE INTERIOR (CONTINUED) THE NATIONAL WILDLIFE REFUGE SYSTEM PROHIBITED ACTS Disturbing Violations: Filming, Photography, and Light...

  20. 50 CFR 27.72 - Audio equipment.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 50 Wildlife and Fisheries 9 2013-10-01 2013-10-01 false Audio equipment. 27.72 Section 27.72 Wildlife and Fisheries UNITED STATES FISH AND WILDLIFE SERVICE, DEPARTMENT OF THE INTERIOR (CONTINUED) THE NATIONAL WILDLIFE REFUGE SYSTEM PROHIBITED ACTS Disturbing Violations: Filming, Photography, and Light...

  1. Audio/Visual Ratios in Commercial Filmstrips.

    ERIC Educational Resources Information Center

    Gulliford, Nancy L.

    Developed by the Westinghouse Electric Corporation, Video Audio Compressed (VIDAC) is a compressed time, variable rate, still picture television system. This technology made it possible for a centralized library of audiovisual materials to be transmitted over a television channel in very short periods of time. In order to establish specifications…

  2. Providing Students with Formative Audio Feedback

    ERIC Educational Resources Information Center

    Brearley, Francis Q.; Cullen, W. Rod

    2012-01-01

    The provision of timely and constructive feedback is increasingly challenging for busy academics. Ensuring effective student engagement with feedback is equally difficult. Increasingly, studies have explored provision of audio recorded feedback to enhance effectiveness and engagement with feedback. Few, if any, of these focus on purely formative…

  3. : Helmholtz machine estimation .

    E-print Network

    : Helmholtz machine density estimation . . : . . . (supervised learning) , (active learning) (query learning) [1, 3]. . (unsupervised learning), . , [5]. . Helmholtz machine , . Helmholtz machine : Helmholtz machine [2] . Helmholtz machine (generative network) (recognition network) . , , . Helmholtz machine (self

  4. Electric machine

    DOEpatents

    El-Refaie, Ayman Mohamed Fawzi (Niskayuna, NY); Reddy, Patel Bhageerath (Madison, WI)

    2012-07-17

    An interior permanent magnet electric machine is disclosed. The interior permanent magnet electric machine comprises a rotor comprising a plurality of radially placed magnets each having a proximal end and a distal end, wherein each magnet comprises a plurality of magnetic segments and at least one magnetic segment towards the distal end comprises a high resistivity magnetic material.

  5. A haptic-inspired audio approach for structural health monitoring decision-making

    NASA Astrophysics Data System (ADS)

    Mao, Zhu; Todd, Michael; Mascareñas, David

    2015-03-01

    Haptics is the field at the interface of human touch (tactile sensation) and classification, whereby tactile feedback is used to train and inform a decision-making process. In structural health monitoring (SHM) applications, haptic devices have been introduced and applied in a simplified laboratory scale scenario, in which nonlinearity, representing the presence of damage, was encoded into a vibratory manual interface. In this paper, the "spirit" of haptics is adopted, but here ultrasonic guided wave scattering information is transformed into audio (rather than tactile) range signals. After sufficient training, the structural damage condition, including occurrence and location, can be identified through the encoded audio waveforms. Different algorithms are employed in this paper to generate the transformed audio signals and the performance of each encoding algorithms is compared, and also compared with standard machine learning classifiers. In the long run, the haptic decision-making is aiming to detect and classify structural damages in a more rigorous environment, and approaching a baseline-free fashion with embedded temperature compensation.

  6. Three-dimensional audio using loudspeakers

    NASA Astrophysics Data System (ADS)

    Gardner, William G.

    1997-12-01

    3-D audio systems, which can surround a listener with sounds at arbitrary locations, are an important part of immersive interfaces. A new approach is presented for implementing 3-D audio using a pair of conventional loudspeakers. The new idea is to use the tracked position of the listener's head to optimize the acoustical presentation, and thus produce a much more realistic illusion over a larger listening area than existing loudspeaker 3-D audio systems. By using a remote head tracker, for instance based on computer vision, an immersive audio environment can be created without donning headphones or other equipment. The general approach to a 3-D audio system is to reconstruct the acoustic pressures at the listener's ears that would result from the natural listening situation to be simulated. To accomplish this using loudspeakers requires that first, the ear signals corresponding to the target scene are synthesized by appropriately encoding directional cues, a process known as 'binaural synthesis,' and second, these signals are delivered to the listener by inverting the transmission paths that exist from the speakers to the listener, a process known as 'crosstalk cancellation.' Existing crosstalk cancellation systems only function at a fixed listening location; when the listener moves away from the equalization zone, the 3-D illusion is lost. Steering the equalization zone to the tracked listener preserves the 3-D illusion over a large listening volume, thus simulating a reconstructed soundfield, and also provides dynamic localization cues by maintaining stationary external sound sources during head motion. This dissertation will discuss the theory, implementation, and testing of a head-tracked loudspeaker 3-D audio system. Crosstalk cancellers that can be steered to the location of a tracked listener will be described. The objective performance of these systems has been evaluated using simulations and acoustical measurements made at the ears of human subjects. Many sound localization experiments were also conducted; the results show that head-tracking both significantly improves localization when the listener is displaced from the ideal listening location, and also enables dynamic localization cues. (Copies available exclusively from MIT Libraries, Rm. 14-0551, Cambridge, MA 02139-4307. Ph. 617-253-5668; Fax 617-253-1690.)

  7. BOOSTED BINARY AUDIO FINGERPRINT BASED ON SPECTRAL SUBBAND MOMENTS Sungwoong Kim and Chang D. Yoo

    E-print Network

    Yoo, Chang D.

    BOOSTED BINARY AUDIO FINGERPRINT BASED ON SPECTRAL SUBBAND MOMENTS Sungwoong Kim and Chang D. Yoo@kaist.ac.kr, cdyoo@ee.kaist.ac.kr ABSTRACT An audio fingerprinting system identifies an audio based on a unique feature vector called the audio fingerprint. The performance of an audio fingerprinting system is directly

  8. Audio feature extraction using probability distribution function

    NASA Astrophysics Data System (ADS)

    Suhaib, A.; Wan, Khairunizam; Aziz, Azri A.; Hazry, D.; Razlan, Zuradzman M.; Shahriman A., B.

    2015-05-01

    Voice recognition has been one of the popular applications in robotic field. It is also known to be recently used for biometric and multimedia information retrieval system. This technology is attained from successive research on audio feature extraction analysis. Probability Distribution Function (PDF) is a statistical method which is usually used as one of the processes in complex feature extraction methods such as GMM and PCA. In this paper, a new method for audio feature extraction is proposed which is by using only PDF as a feature extraction method itself for speech analysis purpose. Certain pre-processing techniques are performed in prior to the proposed feature extraction method. Subsequently, the PDF result values for each frame of sampled voice signals obtained from certain numbers of individuals are plotted. From the experimental results obtained, it can be seen visually from the plotted data that each individuals' voice has comparable PDF values and shapes.

  9. Perceptually controlled doping for audio source separation

    NASA Astrophysics Data System (ADS)

    Mahé, Gaël; Nadalin, Everton Z.; Suyama, Ricardo; Romano, João MT

    2014-12-01

    The separation of an underdetermined audio mixture can be performed through sparse component analysis (SCA) that relies however on the strong hypothesis that source signals are sparse in some domain. To overcome this difficulty in the case where the original sources are available before the mixing process, the informed source separation (ISS) embeds in the mixture a watermark, which information can help a further separation. Though powerful, this technique is generally specific to a particular mixing setup and may be compromised by an additional bitrate compression stage. Thus, instead of watermarking, we propose a `doping' method that makes the time-frequency representation of each source more sparse, while preserving its audio quality. This method is based on an iterative decrease of the distance between the distribution of the signal and a target sparse distribution, under a perceptual constraint. We aim to show that the proposed approach is robust to audio coding and that the use of the sparsified signals improves the source separation, in comparison with the original sources. In this work, the analysis is made only in instantaneous mixtures and focused on voice sources.

  10. Synthecology: sound use of audio in teleimmersion

    NASA Astrophysics Data System (ADS)

    Baum, Geoffrey; Gotsis, Marientina; Chang, Benjamin; Drinkwater, Robb; St. Clair, Dan

    2006-02-01

    This paper examines historical audio applications used to provide real-time immersive sound for CAVE TM environments and discusses their relative strengths and weaknesses. We examine and explain issues of providing spatialized sound immersion in real-time virtual environments (VEs), some problems with currently used sound servers, and a set of requirements for an 'ideal' sound server. We present the initial configuration of a new cross-platform sound server solution using open source software and the Open Sound Control (OSC) specification for the creation of real-time spatialized audio with CAVE applications, specifically Ygdrasil (Yg) environments. The application, aNother Sound Server (NSS) establishes an application interface (API) using OSC, a logical server layer implemented in Python, and an audio engine using SuperCollider (SC). We discuss spatialization implementation and other features. Finally, we document the Synthecology project which premiered at WIRED NEXTFEST 2005 and was the first VE to use NSS. We also discuss various techniques that enhance presence in networked VEs, as well as possible and planned extensions of NSS.

  11. Audio processing technology for law enforcement

    NASA Astrophysics Data System (ADS)

    Walter, Sharon M.; Cofano, Maria; Ratley, Roy J.

    1999-01-01

    The Air Force Research Laboratory Multi-Sensor Exploitation Branch (AFRL/IFEC) has been a Department of Defense leader in research and development (R&D) in speech and audio processing for over 25 years. Their primary thrust in these R&D areas has focused on developing technology to improve the collection, handling, identification, and intelligibility of military communication signals. The National Law Enforcement and Corrections Technology Center for the Northeast (NLECTC-NE) is collocated with the AFRL Rome Research Sited at the Griffiss Technology park in upstate New York. The NLECTC-NE supports sixteen (16) states in the northeast sector of the United States, and is funded and supported by the National Institute of Justice (NIJ). Since the inception of the NLECTC-NE in 1995, the AFRL Rome Research Site has expanded the military applications of their expertise to address law enforcement and corrections requirements. AFRL/IFEC's speech and audio processing technology is unique and particularly appropriate for application to law enforcement requirements. It addresses the similar military need for time-critical decisions and actions, operation within noisy environments, and use by uncooperative speakers in tactical, real-time applications. Audio and speech processing technology for both application domains must also often deal with short utterance communications (less than five seconds of speech) and transmission-to-transmission channel variability.

  12. Capacity-optimized mp2 audio watermarking

    NASA Astrophysics Data System (ADS)

    Steinebach, Martin; Dittmann, Jana

    2003-06-01

    Today a number of audio watermarking algorithms have been proposed, some of them at a quality making them suitable for commercial applications. The focus of most of these algorithms is copyright protection. Therefore, transparency and robustness are the most discussed and optimised parameters. But other applications for audio watermarking can also be identified stressing other parameters like complexity or payload. In our paper, we introduce a new mp2 audio watermarking algorithm optimised for high payload. Our algorithm uses the scale factors of an mp2 file for watermark embedding. They are grouped and masked based on a pseudo-random pattern generated from a secret key. In each group, we embed one bit. Depending on the bit to embed, we change the scale factors by adding 1 where necessary until it includes either more even or uneven scale factors. An uneven group has a 1 embedded, an even group a 0. The same rule is later applied to detect the watermark. The group size can be increased or decreased for transparency/payload trade-off. We embed 160 bits or more in an mp2 file per second without reducing perceived quality. As an application example, we introduce a prototypic Karaoke system displaying song lyrics embedded as a watermark.

  13. Machine therapy

    E-print Network

    Dobson, Kelly E. (Kelly Elizabeth), 1970-

    2007-01-01

    Machine Therapy is a new practice combining art, design, psychoanalysis, and engineering work in ways that access and reveal the vital, though often unnoticed, relevance of people's interactions and relationships with ...

  14. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2013 CFR

    2013-04-01

    ... false Graphic, image, audio and video material. 232.304 Section 232...232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio or video material in a document...

  15. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2014 CFR

    2014-04-01

    ... false Graphic, image, audio and video material. 232.304 Section 232...232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio or video material in a document...

  16. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2012 CFR

    2012-04-01

    ... false Graphic, image, audio and video material. 232.304 Section 232...232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio or video material in a document...

  17. IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 11, NO. 4, JULY 2003 365 High-Fidelity Multichannel Audio Coding With

    E-print Network

    Kuo, C.-C. "Jay"

    -organize the bitstream so that the multi- channel audio bitstream has a quality scalable property when it is transmitted coding (AAC), Karhunen-Loève transform (KLT), MPEG, multichannel audio, quality scalable audio. I of providing transparent audio quality at 384 kbit/sec for 5.1 channels [1]. AAC is cur- rently the most

  18. Combination of Audio and Lyrics Features for Genre Classification in Digital Audio Collections

    E-print Network

    Rauber,Andreas

    to differen- tiate between acoustic styles. For many examples or even genres this is true, for instance `Hip-Hop solely to its sound. Yet, a strong focus is put on audio based feature sets throughout the music information retrieval community, as music perception itself is based on sonic characteristics to a large

  19. Workout Machine

    NASA Technical Reports Server (NTRS)

    1995-01-01

    The Orbotron is a tri-axle exercise machine patterned after a NASA training simulator for astronaut orientation in the microgravity of space. It has three orbiting rings corresponding to roll, pitch and yaw. The user is in the middle of the inner ring with the stomach remaining in the center of all axes, eliminating dizziness. Human power starts the rings spinning, unlike the NASA air-powered system. Marketed by Fantasy Factory (formerly Orbotron, Inc.), the machine can improve aerobic capacity, strength and endurance in five to seven minute workouts.

  20. Machine LearningMachine Learning Stephen Scott

    E-print Network

    Scott, Stephen D.

    Machine LearningMachine Learning Stephen Scott Associate Professor Dept. of Computer Science;1/21/2004 Stephen Scott, Univ. of Nebraska 2 What is Machine Learning?What is Machine Learning? Building machines;1/21/2004 Stephen Scott, Univ. of Nebraska 3 What is Learning?What is Learning? Many different answers, depending

  1. Random machines and Human mind Random machines

    E-print Network

    Perronnin, Florence

    Random machines and Human mind Random machines How algorithms emulate randomness Jean-Marc Vincent1;Random machines and Human mind Outline of the lecture 1 Random machines Why generate random numbers ? Random machines Pseudo-random generators 2 and Human mind Randomness detection Generate randomness 2 / 36

  2. DEVELOPMENT AND EVALUATION OF A SYSTEM FOR WEARABLE AUDIO NAVIGATION

    E-print Network

    DEVELOPMENT AND EVALUATION OF A SYSTEM FOR WEARABLE AUDIO NAVIGATION Bruce N. Walker and Jeffrey If it is not possible to use vision when navigating through one's surroundings, moving safely and effectively becomes much harder. In such cases, non-speech audio cues can serve as navigation beacons, as well as denote

  3. Making the Most of Audio. Technology in Language Learning Series.

    ERIC Educational Resources Information Center

    Barley, Anthony

    Prepared for practicing language teachers, this book's aim is to help them make the most of audio, a readily accessible resource. The book shows, with the help of numerous practical examples, how a range of language skills can be developed. Most examples are in French. Chapters cover the following information: (1) making the most of audio (e.g.,…

  4. Development and Exploration of a Timbre Space Representation of Audio

    E-print Network

    Brewster, Stephen

    Development and Exploration of a Timbre Space Representation of Audio Craig Andrew Nicol Submitted the design possibilities allowed by the new system. The core of this novel audio design system is the timbre in the timbre space and this path can be manipulated geometrically. Several timbre spaces are compared

  5. Single source noise reduction of received HF audio: experimental study

    NASA Astrophysics Data System (ADS)

    Campbell, Eric C.; Alva, Carlos O.

    2014-05-01

    This paper visits the application of single-source noise reduction on received audio over a HF channel. The noise reduction algorithm is typically used in vocoder noise processing at the transmitter before encoding. This study presents the results of the algorithm effects by objectively measuring audio quality through the use of industry standard PESQ analysis.

  6. Automatic Identification of Music Works through Audio Matching

    E-print Network

    Miotto, Riccardo

    and dissemination. Moreover, music is extensively used as the background of commercials, television shows, and news stories. The automatic identification of music works employed as audio background may be useful for usersAutomatic Identification of Music Works through Audio Matching Riccardo Miotto and Nicola Orio

  7. SMC SAMPLERS FOR MULTIRESOLUTION AUDIO SEQUENCE ALIGNMENT Dogac Basaran

    E-print Network

    Cemgil, A. Taylan

    Sampler 1. INTRODUCTION Audio alignment or fingerprinting is defined in the literature as matching. There are several audio fingerprinting methodologies with high matching per- formance [2]-[7]. In [1], we viewed in different fields such as genetics where DNA strands are assembled from shorter sequences [10] and image

  8. INTRAMODAL AND INTERMODAL FUSION FOR AUDIO-VISUAL BIOMETRIC AUTHENTICATION

    E-print Network

    Mak, Man-Wai

    INTRAMODAL AND INTERMODAL FUSION FOR AUDIO-VISUAL BIOMETRIC AUTHENTICATION Ming-Cheung Cheung for decision making. This two- level fusion approach was applied to audio-visual biometric au- thentication rate reduction of up to 83%. 1. INTRODUCTION Various biometric researches have suggested that no single

  9. Using Audio Books to Improve Reading and Academic Performance

    ERIC Educational Resources Information Center

    Montgomery, Joel R.

    2009-01-01

    This article highlights significant research about what below grade-level reading means in middle school classrooms and suggests a tested approach to improve reading comprehension levels significantly by using audio books. The use of these audio books can improve reading and academic performance for both English language learners (ELLs) and for…

  10. How Text and Audio Chat Change the Online Video Experience

    E-print Network

    Kiesler, Sara

    audio chat preferred it to text chat for talking with others while watching videos with their friendsHow Text and Audio Chat Change the Online Video Experience Justin D. Weisz Computer Science kiesler@cs.cmu.edu ABSTRACT Many online video sites provide a text chat feature so viewers can chat

  11. The Practical Audio-Visual Handbook for Teachers.

    ERIC Educational Resources Information Center

    Scuorzo, Herbert E.

    The use of audio/visual media as an aid to instruction is a common practice in today's classroom. Most teachers, however, have little or no formal training in this field and rarely a knowledgeable coordinator to help them. "The Practical Audio-Visual Handbook for Teachers" discusses the types and mechanics of many of these media forms and proposes…

  12. Audio Design: Creating Multi-sensory Images for the Mind.

    ERIC Educational Resources Information Center

    Ferrington, Gary

    1994-01-01

    Explores the concept of "theater of the mind" and discusses design factors in creating audio works that effectively stimulate mental pictures, including: narrative format in audio scripting; qualities of voice; use of concrete language; music; noise versus silence; and the creation of the illusion of space using monaural, stereophonic, and…

  13. Hijacking Power and Bandwidth from the Mobile Phone's Audio Interface

    E-print Network

    Cafarella, Michael J.

    Hijacking Power and Bandwidth from the Mobile Phone's Audio Interface Ye-Sheng Kuo, Sonal Verma the mobile phone with a low-cost, open interface that can parasitically power external peripherals headset audio port. This interface, called HiJack, allows the mobile phone to easily integrate

  14. Some Characteristics of Audio Description and the Corresponding Moving Image.

    ERIC Educational Resources Information Center

    Turner, James M.

    1998-01-01

    This research is concerned with reusing texts produced by audio describers as a source for automatically deriving shot-level indexing for film and video products. Results reinforce the notion that audio description is not sufficient on its own as a source for generating an index to the image, but it is valuable because it describes what is going…

  15. Technical Evaluation Report. 65. Video-Conferencing with Audio Software

    ERIC Educational Resources Information Center

    Baggaley, Jon; Klaas, Jim

    2006-01-01

    An online conference is illustrated using the format of a TV talk show. The conference combined live audio discussion with visual images spontaneously selected by the moderator in the manner of a TV control-room director. A combination of inexpensive online collaborative tools was used for the event, based on the browser-based audio-conferencing…

  16. Use of Video and Audio Texts in EFL Listening Test

    ERIC Educational Resources Information Center

    Basal, Ahmet; Gülözer, Kaine; Demir, Ibrahim

    2015-01-01

    The study aims to discover whether audio or video modality in a listening test is more beneficial to test takers. In this study, the posttest-only control group design was utilized and quantitative data were collected in order to measure participant performances concerning two types of modality (audio or video) in a listening test. The…

  17. INDEXING SPOKEN AUDIO BY LSA AND SOMS Mikko Kurimo

    E-print Network

    Kurimo, Mikko

    INDEXING SPOKEN AUDIO BY LSA AND SOMS Mikko Kurimo Until June 2000 at IDIAP CP-592, Rue du Simplon. For video annotations and indexing the information in audio is a very important component. Many video) for the video. The de#28;nition of suitable index terms just based on video image analysis is much less straight

  18. MPEG-2 Codecs DVD Quality Video, Audio, and Data Networking

    E-print Network

    Ulm, Universität

    of video/audio products are broadcast quality MPEG-2 networked appliances that deliver one-way or twoMPEG-2 Codecs DVD Quality Video, Audio, and Data Networking The VBrickTM 4000, 5000, 6000 series over an IP or ATM network, and decode and display the video. Applications · Interactive

  19. Selected Audio-Visual Materials for Consumer Education. [New Version.

    ERIC Educational Resources Information Center

    Johnston, William L.

    Ninety-two films, filmstrips, multi-media kits, slides, and audio cassettes, produced between 1964 and 1974, are listed in this selective annotated bibliography on consumer education. The major portion of the bibliography is devoted to films and filmstrips. The main topics of the audio-visual materials include purchasing, advertising, money…

  20. Recent Audio-Visual Materials on the Soviet Union.

    ERIC Educational Resources Information Center

    Clarke, Edith Campbell

    1981-01-01

    Identifies and describes audio-visual materials (films, filmstrips, and audio cassette tapes) about the Soviet Union which have been produced since 1977. For each entry, information is presented on title, time required, date of release, cost (purchase and rental), and an abstract. (DB)

  1. The Effect of Audio and Animation in Multimedia Instruction

    ERIC Educational Resources Information Center

    Koroghlanian, Carol; Klein, James D.

    2004-01-01

    This study investigated the effects of audio, animation, and spatial ability in a multimedia computer program for high school biology. Participants completed a multimedia program that presented content by way of text or audio with lean text. In addition, several instructional sequences were presented either with static illustrations or animations.…

  2. Audio Podcasting in a Tablet PC-Enhanced Biochemistry Course

    ERIC Educational Resources Information Center

    Lyles, Heather; Robertson, Brian; Mangino, Michael; Cox, James R.

    2007-01-01

    This report describes the effects of making audio podcasts of all lectures in a large, basic biochemistry course promptly available to students. The audio podcasts complement a previously described approach in which a tablet PC is used to annotate PowerPoint slides with digital ink to produce electronic notes that can be archived. The fundamentals…

  3. Tune in the Net with RealAudio.

    ERIC Educational Resources Information Center

    Buchanan, Larry

    1997-01-01

    Describes how to connect to the RealAudio Web site to download a player that provides sound from Web pages to the computer through streaming technology. Explains hardware and software requirements and provides addresses for other RealAudio Web sites are provided, including weather information and current news. (LRW)

  4. Social Visualization: Exploring Text, Audio, and Video Interaction

    E-print Network

    Karahalios, Karrie G.

    Social Visualization: Exploring Text, Audio, and Video Interaction Abstract In this workshop, we address the importance and uses of social visualizations. In particular, we explore visualizations of text, audio, and visual interaction data to uncover social connections and interaction patterns in online

  5. PHASE SPECTRUM PREDICTION OF AUDIO SIGNALS Ali Bahrami Rad

    E-print Network

    Virtanen, Tuomas

    PHASE SPECTRUM PREDICTION OF AUDIO SIGNALS Ali Bahrami Rad , Tuomas Virtanen Department of Signal.bahrami.rad@aalto.fi, tuomas.virtanen@tut.fi ABSTRACT Modeling the phases of audio signals has received significantly less and neural networks to predict phases from the neighboring points only in the phase spectrum. The simulation

  6. A Low-Cost Audio Prescription Labeling System Using RFID for Thai Visually-Impaired People.

    PubMed

    Lertwiriyaprapa, Titipong; Fakkheow, Pirapong

    2015-01-01

    This research aims to develop a low-cost audio prescription labeling (APL) system for visually-impaired people by using the RFID system. The developed APL system includes the APL machine and APL software. The APL machine is for visually-impaired people while APL software allows caregivers to record all important information into the APL machine. The main objective of the development of the APL machine is to reduce costs and size by designing all of the electronic devices to fit into one print circuit board. Also, it is designed so that it is easy to use and can become an electronic aid for daily living. The developed APL software is based on Java and MySQL, both of which can operate on various operating platforms and are easy to develop as commercial software. The developed APL system was first evaluated by 5 experts. The APL system was also evaluated by 50 actual visually-impaired people (30 elders and 20 blind individuals) and 20 caregivers, pharmacists and nurses. After using the APL system, evaluations were carried out, and it can be concluded from the evaluation results that this proposed APL system can be effectively used for helping visually-impaired people in terms of self-medication. PMID:26427743

  7. and machines, on-board autonomy, human control, and ground automation.

    E-print Network

    Stone, Peter

    contamination the payload can take. Another example is thermal constraints, such as maximum solar point timeand machines, on-board autonomy, human control, and ground automation. CHASER is comprised of three. The second canister contains the flight computer for the payload as well as the 2 GB Digital Audio Tape (DAT

  8. Wacky Machines

    ERIC Educational Resources Information Center

    Fendrich, Jean

    2002-01-01

    Collectors everywhere know that local antique shops and flea markets are treasure troves just waiting to be plundered. Science teachers might take a hint from these hobbyists, for the next community yard sale might be a repository of old, quirky items that are just the things to get students thinking about simple machines. By introducing some…

  9. Horatio Audio-Describes Shakespeare's "Hamlet": Blind and Low-Vision Theatre-Goers Evaluate an Unconventional Audio Description Strategy

    ERIC Educational Resources Information Center

    Udo, J. P.; Acevedo, B.; Fels, D. I.

    2010-01-01

    Audio description (AD) has been introduced as one solution for providing people who are blind or have low vision with access to live theatre, film and television content. However, there is little research to inform the process, user preferences and presentation style. We present a study of a single live audio-described performance of Hart House…

  10. Machine Learning

    NASA Astrophysics Data System (ADS)

    Hoffmann, Achim; Mahidadia, Ashesh

    The purpose of this chapter is to present fundamental ideas and techniques of machine learning suitable for the field of this book, i.e., for automated scientific discovery. The chapter focuses on those symbolic machine learning methods, which produce results that are suitable to be interpreted and understood by humans. This is particularly important in the context of automated scientific discovery as the scientific theories to be produced by machines are usually meant to be interpreted by humans. This chapter contains some of the most influential ideas and concepts in machine learning research to give the reader a basic insight into the field. After the introduction in Sect. 1, general ideas of how learning problems can be framed are given in Sect. 2. The section provides useful perspectives to better understand what learning algorithms actually do. Section 3 presents the Version space model which is an early learning algorithm as well as a conceptual framework, that provides important insight into the general mechanisms behind most learning algorithms. In section 4, a family of learning algorithms, the AQ family for learning classification rules is presented. The AQ family belongs to the early approaches in machine learning. The next, Sect. 5 presents the basic principles of decision tree learners. Decision tree learners belong to the most influential class of inductive learning algorithms today. Finally, a more recent group of learning systems are presented in Sect. 6, which learn relational concepts within the framework of logic programming. This is a particularly interesting group of learning systems since the framework allows also to incorporate background knowledge which may assist in generalisation. Section 7 discusses Association Rules - a technique that comes from the related field of Data mining. Section 8 presents the basic idea of the Naive Bayesian Classifier. While this is a very popular learning technique, the learning result is not well suited for human comprehension as it is essentially a large collection of probability values. In Sect. 9, we present a generic method for improving accuracy of a given learner by generatingmultiple classifiers using variations of the training data. While this works well in most cases, the resulting classifiers have significantly increased complexity and, hence, tend to destroy the human readability of the learning result that a single learner may produce. Section 10 contains a summary, mentions briefly other techniques not discussed in this chapter and presents outlook on the potential of machine learning in the future.

  11. Fullerene Machines

    NASA Technical Reports Server (NTRS)

    Globus, Al; Saini, Subhash (Technical Monitor)

    1998-01-01

    Fullerenes possess remarkable properties and many investigators have examined the mechanical, electronic and other characteristics of carbon SP2 systems in some detail. In addition, C-60 can be functionalized with many classes of molecular fragments and we may expect the caps of carbon nanotubes to have a similar chemistry. Finally, carbon nanotubes have been attached to t he end of scanning probe microscope (Spill) tips. Spills can be manipulated with sub-angstrom accuracy. Together, these investigations suggest that complex molecular machines made of fullerenes may someday be created and manipulated with very high accuracy. We have studied some such systems computationally (primarily functionalized carbon nanotube gears and computer components). If such machines can be combined appropriately, a class of materials may be created that can sense their environment, calculate a response, and act. The implications of such hypothetical materials are substantial.

  12. Fullerene Machines

    NASA Technical Reports Server (NTRS)

    Globus, Al; Saini, Subhash

    1998-01-01

    Recent computational efforts at NASA Ames Research Center and computation and experiment elsewhere suggest that a nanotechnology of machine phase functionalized fullerenes may be synthetically accessible and of great interest. We have computationally demonstrated that molecular gears fashioned from (14,0) single-walled carbon nanotubes and benzyne teeth should operate well at 50-100 gigahertz. Preliminary results suggest that these gears can be cooled by a helium atmosphere and a laser motor can power fullerene gears if a positive and negative charge have been added to form a dipole. In addition, we have unproven concepts based on experimental and computational evidence for support structures, computer control, a system architecture, a variety of components, and manufacture. Combining fullerene machines with the remarkable mechanical properties of carbon nanotubes, there is some reason to believe that a focused effort to develop fullerene nanotechnology could yield materials with tremendous properties.

  13. Departement Elektrotechniek ESAT-STADIUS/TR 13-35 Declipping of audio signals using perceptual

    E-print Network

    as declipping, is important to achieve an improved level of audio quality in many audio applications evaluation exper- iments reveal a significant audio quality increase for the proposed PCS-based declippingKU Leuven Departement Elektrotechniek ESAT-STADIUS/TR 13-35 Declipping of audio signals using

  14. Self-content-based audio inpainting Yuval Bahat a,n

    E-print Network

    Elad, Michael

    ) mechanism, to counter the degradation in audio quality, by filling in for the missing audio data, usingSelf-content-based audio inpainting Yuval Bahat a,n , Yoav Y. Schechner a , Michael Elad b recorded audio of the same user. Our approach also harnesses statistical priors and data inpainting

  15. Adaptive Rate Control for Real-time Packet Audio Based on Loss Prediction

    E-print Network

    Wang, Yongge

    and on-line audio quality assessment for real-time packet audio. The proposed framework determines quality under network constraints. Quality of an audio communication is highly sensitive to packet loss [3 to changing bandwidth and delay by changing the optimal codec combination, while maintaining the audio quality

  16. Waveform Analysis for High-Quality Loop-Based Audio Distribution Stuart Cunningham

    E-print Network

    Davies, John N.

    Waveform Analysis for High-Quality Loop-Based Audio Distribution Stuart Cunningham Centre using compression procedures, but the quality of the audio suffers, to the detriment of the reproduced audio, and thus, the listening experience. Though acceptable audio quality can be achieved using

  17. Characterization of LMR Analog FM Audio Quality Using PL Tone Analysis

    E-print Network

    Ellingson, Steven W.

    Characterization of LMR Analog FM Audio Quality Using PL Tone Analysis Akshay Kumar Steven for characterizing the land mobile radio (LMR) analog FM audio quality using Private Line (PL) tone analysis of the audio quality in order to accurately determine in-situ RF signal to noise ratio (SNR). The audio signal

  18. FINE-GRAIN SCALABLE AUDIO CODING BASED ON ENVELOPE RESTORATION AND THE SPIHT ALGORITHM

    E-print Network

    Lübeck, Universität zu

    the perceptual audio quality notably. Index Terms-- progressive audio compression, noise shaping, embedded coding perceptual transparent audio quality at fixed target bitrates between 48-128 kbps (kbits per sec- ondFINE-GRAIN SCALABLE AUDIO CODING BASED ON ENVELOPE RESTORATION AND THE SPIHT ALGORITHM Heiko Hansen

  19. Frequency-Based Coloring of the Audio Waveform Display Stephen V. Rice

    E-print Network

    Rice, Stephen V.

    Frequency-Based Coloring of the Audio Waveform Display Stephen V. Rice The University of Mississippi rice@cs.olemiss.edu The audio waveform display provides the visual focus in audio-editing systems. This technique was invented by the author, S. V. Rice. A sequence of digital audio samples is partitioned

  20. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2011 CFR

    2011-04-01

    ... 17 Commodity and Securities Exchanges 2 2011-04-01 2011-04-01 false Graphic, image, audio and... Submissions § 232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio..., image, audio or video material is presented in the delivered version, or they may be listed in...

  1. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2010 CFR

    2010-04-01

    ... 17 Commodity and Securities Exchanges 2 2010-04-01 2010-04-01 false Graphic, image, audio and... Submissions § 232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio..., image, audio or video material is presented in the delivered version, or they may be listed in...

  2. Induction machine

    DOEpatents

    Owen, Whitney H. (Ogden, UT)

    1980-01-01

    A polyphase rotary induction machine for use as a motor or generator utilizing a single rotor assembly having two series connected sets of rotor windings, a first stator winding disposed around the first rotor winding and means for controlling the current induced in one set of the rotor windings compared to the current induced in the other set of the rotor windings. The rotor windings may be wound rotor windings or squirrel cage windings.

  3. Effects of aging on audio-visual speech integration.

    PubMed

    Huyse, Aurélie; Leybaert, Jacqueline; Berthommier, Frédéric

    2014-10-01

    This study investigated the impact of aging on audio-visual speech integration. A syllable identification task was presented in auditory-only, visual-only, and audio-visual congruent and incongruent conditions. Visual cues were either degraded or unmodified. Stimuli were embedded in stationary noise alternating with modulated noise. Fifteen young adults and 15 older adults participated in this study. Results showed that older adults had preserved lipreading abilities when the visual input was clear but not when it was degraded. The impact of aging on audio-visual integration also depended on the quality of the visual cues. In the visual clear condition, the audio-visual gain was similar in both groups and analyses in the framework of the fuzzy-logical model of perception confirmed that older adults did not differ from younger adults in their audio-visual integration abilities. In the visual reduction condition, the audio-visual gain was reduced in the older group, but only when the noise was stationary, suggesting that older participants could compensate for the loss of lipreading abilities by using the auditory information available in the valleys of the noise. The fuzzy-logical model of perception confirmed the significant impact of aging on audio-visual integration by showing an increased weight of audition in the older group. PMID:25324091

  4. MPEG-4 low-delay general audio coding

    NASA Astrophysics Data System (ADS)

    Sporer, Thomas; Grill, Bernhard; Herre, Juergen

    2001-07-01

    Traditionally, speech coding for communication purposes and perceptual audio coding have been separate worlds. On one hand, speech coders provide acceptable speech quality at very low data rates and low delays which are suitable for two-way communication applications, such as Voice over IP (VoIP) or teleconferencing. Due to the underlying coding paradigm, however, such coders do not perform well for non-speech signals (e.g.~music and environmental noise). Furthermore, the sound quality and naturalness is severely limited by the fact that most coders are working in narrow-band mode, i.e. with a bandwidth below 4 kHz. On the other hand, perceptual audio codecs provide excellent subjective audio quality for a broad range of signals including speech at bit rates down to 16 kbit/s. The delay of such a coder/decoder chain, however, usually exceeds 200 ms at very low data rates and in this way is not acceptable for interactive two-way communication. This paper describes a coding scheme which is designed to combine the advantages of perceptual audio coding with the low delay necessary for two-way communication. The codec was standardized within MPEG-4 Version 2 Audio under the work item ``Low Delay Audio Coding'' and is derived from the ISO/MPEG-2/4 Advanced Audio Coding (AAC) algorithm. The algorithm provides modes operating at algorithmic delay as low as 20 ms and is equipped to handle all full-bandwidth high-quality audio signals, both in monophonic, stereophonic and even multi-channel format. Despite of the low algorithmic delay, the codec delivers better audio quality than MPEG-1 Layer-3 (MP3) at the same bit rate. The paper also addresses issues pertaining to the integration of the coder into H.32x and SDP applications.

  5. Musical examination to bridge audio data and sheet music

    NASA Astrophysics Data System (ADS)

    Pan, Xunyu; Cross, Timothy J.; Xiao, Liangliang; Hei, Xiali

    2015-03-01

    The digitalization of audio is commonly implemented for the purpose of convenient storage and transmission of music and songs in today's digital age. Analyzing digital audio for an insightful look at a specific musical characteristic, however, can be quite challenging for various types of applications. Many existing musical analysis techniques can examine a particular piece of audio data. For example, the frequency of digital sound can be easily read and identified at a specific section in an audio file. Based on this information, we could determine the musical note being played at that instant, but what if you want to see a list of all the notes played in a song? While most existing methods help to provide information about a single piece of the audio data at a time, few of them can analyze the available audio file on a larger scale. The research conducted in this work considers how to further utilize the examination of audio data by storing more information from the original audio file. In practice, we develop a novel musical analysis system Musicians Aid to process musical representation and examination of audio data. Musicians Aid solves the previous problem by storing and analyzing the audio information as it reads it rather than tossing it aside. The system can provide professional musicians with an insightful look at the music they created and advance their understanding of their work. Amateur musicians could also benefit from using it solely for the purpose of obtaining feedback about a song they were attempting to play. By comparing our system's interpretation of traditional sheet music with their own playing, a musician could ensure what they played was correct. More specifically, the system could show them exactly where they went wrong and how to adjust their mistakes. In addition, the application could be extended over the Internet to allow users to play music with one another and then review the audio data they produced. This would be particularly useful for teaching music lessons on the web. The developed system is evaluated with songs played with guitar, keyboard, violin, and other popular musical instruments (primarily electronic or stringed instruments). The Musicians Aid system is successful at both representing and analyzing audio data and it is also powerful in assisting individuals interested in learning and understanding music.

  6. Noise-Canceling Helmet Audio System

    NASA Technical Reports Server (NTRS)

    Seibert, Marc A.; Culotta, Anthony J.

    2007-01-01

    A prototype helmet audio system has been developed to improve voice communication for the wearer in a noisy environment. The system was originally intended to be used in a space suit, wherein noise generated by airflow of the spacesuit life-support system can make it difficult for remote listeners to understand the astronaut s speech and can interfere with the astronaut s attempt to issue vocal commands to a voice-controlled robot. The system could be adapted to terrestrial use in helmets of protective suits that are typically worn in noisy settings: examples include biohazard, fire, rescue, and diving suits. The system (see figure) includes an array of microphones and small loudspeakers mounted at fixed positions in a helmet, amplifiers and signal-routing circuitry, and a commercial digital signal processor (DSP). Notwithstanding the fixed positions of the microphones and loudspeakers, the system can accommodate itself to any normal motion of the wearer s head within the helmet. The system operates in conjunction with a radio transceiver. An audio signal arriving via the transceiver intended to be heard by the wearer is adjusted in volume and otherwise conditioned and sent to the loudspeakers. The wearer s speech is collected by the microphones, the outputs of which are logically combined (phased) so as to form a microphone- array directional sensitivity pattern that discriminates in favor of sounds coming from vicinity of the wearer s mouth and against sounds coming from elsewhere. In the DSP, digitized samples of the microphone outputs are processed to filter out airflow noise and to eliminate feedback from the loudspeakers to the microphones. The resulting conditioned version of the wearer s speech signal is sent to the transceiver.

  7. The power of digital audio in interactive instruction: An unexploited medium

    SciTech Connect

    Pratt, J.; Trainor, M.

    1989-01-01

    Widespread use of audio in computer-based training (CBT) occurred with the advent of the interactive videodisc technology. This paper discusses the alternative of digital audio, which, unlike videodisc audio, enables one to rapidly revise the audio used in the CBT and which may be used in nonvideo CBT applications as well. We also discuss techniques used in audio script writing, editing, and production. Results from evaluations indicate a high degree of user satisfaction. 4 refs.

  8. Proceedings of ICAD 05-Eleventh Meeting of the International Conference on Auditory Display, Limerick, Ireland, July 6-9, 2005 MOBILE AUDIO DESIGNS MONKEY: AN AUDIO AUGMENTED

    E-print Network

    , Limerick, Ireland, July 6-9, 2005 ICAD05-1 MOBILE AUDIO DESIGNS MONKEY: AN AUDIO AUGMENTED REALITY DESIGNER) Monkey was developed. MAD Monkey was developed using the standard User Centered Design process

  9. TEMPO machine

    SciTech Connect

    Rohwein, G.J.; Lancaster, K.T.; Lawson, R.N.

    1986-06-01

    TEMPO is a transformer powered megavolt pulse generator with an output pulse of 100 ns duration. The machine was designed for burst mode operation at pulse repetition rates up to 10 Hz with minimum pulse-to-pulse voltage variations. To meet the requirement for pulse duration a nd a 20-..omega.. output impedance within reasonable size constraints, the pulse forming transmission line was designed as two parallel water-insulated, strip-type Blumleins. Stray capacitance and electric fields along the edges of the line elements were controlled by lining the tank with plastic sheet.

  10. Applying Spatial Audio to Human Interfaces: 25 Years of NASA Experience

    NASA Technical Reports Server (NTRS)

    Begault, Durand R.; Wenzel, Elizabeth M.; Godfrey, Martine; Miller, Joel D.; Anderson, Mark R.

    2010-01-01

    From the perspective of human factors engineering, the inclusion of spatial audio within a human-machine interface is advantageous from several perspectives. Demonstrated benefits include the ability to monitor multiple streams of speech and non-speech warning tones using a cocktail party advantage, and for aurally-guided visual search. Other potential benefits include the spatial coordination and interaction of multimodal events, and evaluation of new communication technologies and alerting systems using virtual simulation. Many of these technologies were developed at NASA Ames Research Center, beginning in 1985. This paper reviews examples and describes the advantages of spatial sound in NASA-related technologies, including space operations, aeronautics, and search and rescue. The work has involved hardware and software development as well as basic and applied research.

  11. COMBINING FEATURES REDUCES HUBNESS IN AUDIO SIMILARITY Arthur Flexer,1

    E-print Network

    Widmer, Gerhard

    . The de facto standard approach to computa- tion of audio similarity is timbre similarity based on para that a non-timbre based parameterization is not prone to hubness; (iii) finally we show how combining timbre

  12. Direct broadcast satellite-audio, portable and mobile reception tradeoffs

    NASA Technical Reports Server (NTRS)

    Golshan, Nasser

    1992-01-01

    This paper reports on the findings of a systems tradeoffs study on direct broadcast satellite-radio (DBS-R). Based on emerging advanced subband and transform audio coding systems, four ranges of bit rates: 16-32 kbps, 48-64 kbps, 96-128 kbps and 196-256 kbps are identified for DBS-R. The corresponding grades of audio quality will be subjectively comparable to AM broadcasting, monophonic FM, stereophonic FM, and CD quality audio, respectively. The satellite EIRP's needed for mobile DBS-R reception in suburban areas are sufficient for portable reception in most single family houses when allowance is made for the higher G/T of portable table-top receivers. As an example, the variation of the space segment cost as a function of frequency, audio quality, coverage capacity, and beam size is explored for a typical DBS-R system.

  13. Audio CAPTCHA for SIP-Based VoIP

    NASA Astrophysics Data System (ADS)

    Soupionis, Yannis; Tountas, George; Gritzalis, Dimitris

    Voice over IP (VoIP) introduces new ways of communication, while utilizing existing data networks to provide inexpensive voice communications worldwide as a promising alternative to the traditional PSTN telephony. SPam over Internet Telephony (SPIT) is one potential source of future annoyance in VoIP. A common way to launch a SPIT attack is the use of an automated procedure (bot), which generates calls and produces audio advertisements. In this paper, our goal is to design appropriate CAPTCHA to fight such bots. We focus on and develop audio CAPTCHA, as the audio format is more suitable for VoIP environments and we implement it in a SIP-based VoIP environment. Furthermore, we suggest and evaluate the specific attributes that audio CAPTCHA should incorporate in order to be effective, and test it against an open source bot implementation.

  14. Audio Segment Retrieval Using a Synthesized HMM Atulya Velivelli

    E-print Network

    Zhai, ChengXiang

    launches in the last century. The audio theme corresponding to a rocket launch is an explosion. The user way of specifying the information need. For example if we have a video about the different rocket

  15. Information Services Audio-Visual Strategy 2015 2018

    E-print Network

    Kent, University of

    Information Services Audio-Visual Strategy 2015 ­ 2018 (Learning and Teaching) Context requirements and through the monitoring of technology advances help shape longer term requirements to support supports the following Information Services strategic aims2 : Environment - We offer modern, technology

  16. 47 CFR 87.483 - Audio visual warning systems.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ...audio visual warning system (AVWS) is a radar-based obstacle avoidance system...obstructions exist. The continuously operating radar calculates the location, direction...obstacle. (a) Radiodetermination (radar) frequencies. Frequencies...

  17. HGS Schedulers for Digital Audio Workstation like Applications

    E-print Network

    Poduval, Karthik Venugopal

    2014-08-31

    Digital Audio Workstation (DAW) applications are real-time applications that have special timing constraints. Hierarchical Group Scheduling (HGS) is a real-time scheduling framework that allows developers implement custom ...

  18. HMM-based Speech Synthesis from Audio Book Data 

    E-print Network

    Haag, Kathrin

    In contrast to hand-crafted speech databases, which contain short out-of-context sentences in fairly unemphatic speech style, audio books contain rich prosody including intonation contours, pitch accents and phrasing patterns, which is a good pre...

  19. Speech and crosstalk detection in multi-channel audio 

    E-print Network

    Wrigley, Stuart N; Brown, Guy J; Wan, Vincent; Renals, Steve

    2005-01-01

    related to the automatic classification of audio into these four classes. The first experiment attempted to optimise a set of acoustic features for use with a Gaussian mixture model (GMM) classifier. A large set of potential acoustic features were...

  20. Virtual environment interaction through 3D audio by blind children.

    PubMed

    Sánchez, J; Lumbreras, M

    1999-01-01

    Interactive software is actively used for learning, cognition, and entertainment purposes. Educational entertainment software is not very popular among blind children because most computer games and electronic toys have interfaces that are only accessible through visual cues. This work applies the concept of interactive hyperstories to blind children. Hyperstories are implemented in a 3D acoustic virtual world. In past studies we have conceptualized a model to design hyperstories. This study illustrates the feasibility of the model. It also provides an introduction to researchers to the field of entertainment software for blind children. As a result, we have designed and field tested AudioDoom, a virtual environment interacted through 3D Audio by blind children. AudioDoom is also a software that enables testing nontrivial interfaces and cognitive tasks with blind children. We explored the construction of cognitive spatial structures in the minds of blind children through audio-based entertainment and spatial sound navigable experiences. Children playing AudioDoom were exposed to first person experiences by exploring highly interactive virtual worlds through the use of 3D aural representations of the space. This experience was structured in several cognitive tasks where they had to build concrete models of their spatial representations constructed through the interaction with AudioDoom by using Legotrade mark blocks. We analyze our preliminary results after testing AudioDoom with Chilean children from a school for blind children. We discuss issues such as interactivity in software without visual cues, the representation of spatial sound navigable experiences, and entertainment software such as computer games for blind children. We also evaluate the feasibility to construct virtual environments through the design of dynamic learning materials with audio cues. PMID:19178246

  1. Tunneling machine

    SciTech Connect

    Snyder, L.L.

    1980-02-19

    A diametrically compact tunneling machine for boring tunnels is disclosed. The machine includes a tubular support frame having a hollow piston mounted therein which is movable from a retracted position in the support frame to an extended position. A drive shaft is rotatably mounted in the hollow piston and carries a cutter head at one end. The hollow piston is restrained against rotational movement relative to the support frame and the drive shaft is constrained against longitudinal movement relative to the hollow piston. A plurality of radially extendible feet project from the support frame to the tunnel wall to grip the tunnel wall during a tunneling operation wherein the hollow piston is driven forwardly so that the cutter head works on the tunnel face. When the hollow piston is fully extended, a plurality of extendible support feet, which are fixed to the rearward and forward ends of the hollow piston, are extended, the radially extendible feet are retracted and the support frame is shifted forwardly by the piston so that a further tunneling operation may be initiated.

  2. IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 20, NO. 10, DECEMBER 2012 2657 Real-Time Perception-Based Clipping of Audio

    E-print Network

    on the perceived audio signal quality. In this paper, we present a novel multidis- ciplinary approach to clipping are provided. From comparative audio quality evaluation experiments, it is concluded that the perception-based clipping algorithm results in significantly higher objective audio quality scores than existing clipping

  3. 2011 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics October 16-19, 2011, New Paltz, NY MODEL BASED MULTIPLE AUDIO SEQUENCE ALIGNMENT

    E-print Network

    Cemgil, A. Taylan

    of the concert with record- ing devices of varying quality. These audio clips each of which are recorded from2011 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics October 16-19, 2011, New Paltz, NY MODEL BASED MULTIPLE AUDIO SEQUENCE ALIGNMENT Dogac¸ Bas¸aran , A. Taylan Cemgil

  4. PREPRINT FOR AN OVERVIEW ON PERCEPTUALLY MOTIVATED AUDIO INDEXING AND CLASSIFICATION, PROC. OF THE IEEE, 2013 1 An overview on Perceptually Motivated Audio

    E-print Network

    Richard, Gaël

    An audio indexing system aims at describing audio content by identifying, labeling or categorizing different acoustic events. Since the resulting audio classification and indexing is meant for direct human by integrating specific knowledge of the human auditory system in the design process to various extent

  5. Course info Machine Learning

    E-print Network

    Shi, Qinfeng "Javen"

    Course info Machine Learning Real life problems Lecture 1: Machine Learning Problem Qinfeng (Javen) Shi 28 July 2014 Intro. to Stats. Machine Learning COMP SCI 4401/7401 Qinfeng (Javen) Shi Lecture 1: Machine Learning Problem #12;Course info Machine Learning Real life problems Table of Contents I 1 Course

  6. Roel Wieringa Machine Intelligence

    E-print Network

    Wieringa, Roel

    Roel Wieringa Machine Intelligence and Explication Rapport nr.IR-129 VRIJE UNIVERSITEITAMSTERDAM Facultei~ Wiskunde en Informatica, De Boelelaan 1081, 1081 HV Amsterdam #12;Machine Intelligence. Machines and explicitness 1.1 Systems, states, and processes 1.2 Machines, Turing machines, computers 1

  7. Sony's Data Discman: A Look at These New Portable Information Machines and What They Mean for CD-ROM Developers.

    ERIC Educational Resources Information Center

    Bonime, Andrew

    1992-01-01

    Describes a portable CD-ROM machine intended for the mass market that provides access to searchable text, graphics, and audio through a user-friendly interface. Six search modes and other system features are reviewed, and electronic texts for the unit are introduced. A table compares features of the two available models. (NRP)

  8. Socially guided machine learning

    E-print Network

    Thomaz, Andrea Lockerd

    2006-01-01

    Social interaction will be key to enabling robots and machines in general to learn new tasks from ordinary people (not experts in robotics or machine learning). Everyday people who need to teach their machines new things ...

  9. Drawing Machines NAEA 2013

    E-print Network

    is a powerful way to introduce kinetic concepts to an arts class: drawing machines drawing machines are kinetic sculptures that make drawings, typically drawing with a short history of drawing machines, leading up to examples of contemporary art

  10. Digital Audio Radio Broadcast Systems Laboratory Testing Nearly Complete

    NASA Technical Reports Server (NTRS)

    2005-01-01

    Radio history continues to be made at the NASA Lewis Research Center with the completion of phase one of the digital audio radio (DAR) testing conducted by the Consumer Electronics Group of the Electronic Industries Association. This satellite, satellite/terrestrial, and terrestrial digital technology will open up new audio broadcasting opportunities both domestically and worldwide. It will significantly improve the current quality of amplitude-modulated/frequency-modulated (AM/FM) radio with a new digitally modulated radio signal and will introduce true compact-disc-quality (CD-quality) sound for the first time. Lewis is hosting the laboratory testing of seven proposed digital audio radio systems and modes. Two of the proposed systems operate in two modes each, making a total of nine systems being tested. The nine systems are divided into the following types of transmission: in-band on-channel (IBOC), in-band adjacent-channel (IBAC), and new bands. The laboratory testing was conducted by the Consumer Electronics Group of the Electronic Industries Association. Subjective assessments of the audio recordings for each of the nine systems was conducted by the Communications Research Center in Ottawa, Canada, under contract to the Electronic Industries Association. The Communications Research Center has the only CCIR-qualified (Consultative Committee for International Radio) audio testing facility in North America. The main goals of the U.S. testing process are to (1) provide technical data to the Federal Communication Commission (FCC) so that it can establish a standard for digital audio receivers and transmitters and (2) provide the receiver and transmitter industries with the proper standards upon which to build their equipment. In addition, the data will be forwarded to the International Telecommunications Union to help in the establishment of international standards for digital audio receivers and transmitters, thus allowing U.S. manufacturers to compete in the world market.

  11. Machine musicianship

    NASA Astrophysics Data System (ADS)

    Rowe, Robert

    2002-05-01

    The training of musicians begins by teaching basic musical concepts, a collection of knowledge commonly known as musicianship. Computer programs designed to implement musical skills (e.g., to make sense of what they hear, perform music expressively, or compose convincing pieces) can similarly benefit from access to a fundamental level of musicianship. Recent research in music cognition, artificial intelligence, and music theory has produced a repertoire of techniques that can make the behavior of computer programs more musical. Many of these were presented in a recently published book/CD-ROM entitled Machine Musicianship. For use in interactive music systems, we are interested in those which are fast enough to run in real time and that need only make reference to the material as it appears in sequence. This talk will review several applications that are able to identify the tonal center of musical material during performance. Beyond this specific task, the design of real-time algorithmic listening through the concurrent operation of several connected analyzers is examined. The presentation includes discussion of a library of C++ objects that can be combined to perform interactive listening and a demonstration of their capability.

  12. Audio-video feature correlation: faces and speech

    NASA Astrophysics Data System (ADS)

    Durand, Gwenael; Montacie, Claude; Caraty, Marie-Jose; Faudemay, Pascal

    1999-08-01

    This paper presents a study of the correlation of features automatically extracted from the audio stream and the video stream of audiovisual documents. In particular, we were interested in finding out whether speech analysis tools could be combined with face detection methods, and to what extend they should be combined. A generic audio signal partitioning algorithm as first used to detect Silence/Noise/Music/Speech segments in a full length movie. A generic object detection method was applied to the keyframes extracted from the movie in order to detect the presence or absence of faces. The correlation between the presence of a face in the keyframes and of the corresponding voice in the audio stream was studied. A third stream, which is the script of the movie, is warped on the speech channel in order to automatically label faces appearing in the keyframes with the name of the corresponding character. We naturally found that extracted audio and video features were related in many cases, and that significant benefits can be obtained from the joint use of audio and video analysis methods.

  13. Audio-visual integration through the parallel visual pathways.

    PubMed

    Kaposvári, Péter; Csete, Gerg?; Bognár, Anna; Csibri, Péter; Tóth, Eszter; Szabó, Nikoletta; Vécsei, László; Sáry, Gyula; Tamás Kincses, Zsigmond

    2015-10-22

    Audio-visual integration has been shown to be present in a wide range of different conditions, some of which are processed through the dorsal, and others through the ventral visual pathway. Whereas neuroimaging studies have revealed integration-related activity in the brain, there has been no imaging study of the possible role of segregated visual streams in audio-visual integration. We set out to determine how the different visual pathways participate in this communication. We investigated how audio-visual integration can be supported through the dorsal and ventral visual pathways during the double flash illusion. Low-contrast and chromatic isoluminant stimuli were used to drive preferably the dorsal and ventral pathways, respectively. In order to identify the anatomical substrates of the audio-visual interaction in the two conditions, the psychophysical results were correlated with the white matter integrity as measured by diffusion tensor imaging.The psychophysiological data revealed a robust double flash illusion in both conditions. A correlation between the psychophysical results and local fractional anisotropy was found in the occipito-parietal white matter in the low-contrast condition, while a similar correlation was found in the infero-temporal white matter in the chromatic isoluminant condition. Our results indicate that both of the parallel visual pathways may play a role in the audio-visual interaction. PMID:26165152

  14. MedlinePlus FAQ: Is audio description available for videos on MedlinePlus?

    MedlinePLUS

    ... audiodescription.html Question: Is audio description available for videos on MedlinePlus? To use the sharing features on ... page, please enable JavaScript. Answer: Audio description of videos helps make the content of videos accessible to ...

  15. 16 CFR 307.8 - Requirements for disclosure in audiovisual and audio advertising.

    Code of Federal Regulations, 2010 CFR

    2010-01-01

    ...disclosure in audiovisual and audio advertising. 307.8 Section 307...HEALTH EDUCATION ACT OF 1986 Advertising Disclosures § 307.8 Requirements...disclosure in audiovisual and audio advertising. In the case of...

  16. An Automatic Commercial Search Application for TV Broadcasting Using Audio Fingerprinting 

    E-print Network

    Song, Yaohua

    2012-10-19

    it because it takes much more storage and computations to process. For the audio stream, we have to extract proper audio features which can represent its characteristics and store them in a database for identification. We choose the Normalized Spectral...

  17. Indexing Audio Documents by using Latent Semantic Analysis and SOM Mikko Kurimo a

    E-print Network

    Kurimo, Mikko

    Indexing Audio Documents by using Latent Semantic Analysis and SOM Mikko Kurimo a a IDIAP CP­592 recalls the relevant parts of audio or video broadcasts based on a query from the user. A prototype system

  18. Machine Shop Lathes.

    ERIC Educational Resources Information Center

    Dunn, James

    This guide, the second in a series of five machine shop curriculum manuals, was designed for use in machine shop courses in Oklahoma. The purpose of the manual is to equip students with basic knowledge and skills that will enable them to enter the machine trade at the machine-operator level. The curriculum is designed so that it can be used in…

  19. Applied machine vision

    SciTech Connect

    Not Available

    1984-01-01

    This book presents the papers given at a conference on robot vision. Topics considered at the conference included the link between fixed and flexible automation, general applications of machine vision, the development of a specification for a machine vision system, machine vision technology, machine vision non-contact gaging, and vision in electronics manufacturing.

  20. Multi-channel spatialization system for audio signals

    NASA Technical Reports Server (NTRS)

    Begault, Durand R. (inventor)

    1995-01-01

    Synthetic head related transfer functions (HRTF's) for imposing reprogramable spatial cues to a plurality of audio input signals included, for example, in multiple narrow-band audio communications signals received simultaneously are generated and stored in interchangeable programmable read only memories (PROM's) which store both head related transfer function impulse response data and source positional information for a plurality of desired virtual source locations. The analog inputs of the audio signals are filtered and converted to digital signals from which synthetic head related transfer functions are generated in the form of linear phase finite impulse response filters. The outputs of the impulse response filters are subsequently reconverted to analog signals, filtered, mixed and fed to a pair of headphones.

  1. Highlight summarization in golf videos using audio signals

    NASA Astrophysics Data System (ADS)

    Kim, Hyoung-Gook; Kim, Jin Young

    2008-01-01

    In this paper, we present an automatic summarization of highlights in golf videos based on audio information alone without video information. The proposed highlight summarization system is carried out based on semantic audio segmentation and detection on action units from audio signals. Studio speech, field speech, music, and applause are segmented by means of sound classification. Swing is detected by the methods of impulse onset detection. Sounds like swing and applause form a complete action unit, while studio speech and music parts are used to anchor the program structure. With the advantage of highly precise detection of applause, highlights are extracted effectively. Our experimental results obtain high classification precision on 18 golf games. It proves that the proposed system is very effective and computationally efficient to apply the technology to embedded consumer electronic devices.

  2. Music Identification System Using MPEG-7 Audio Signature Descriptors

    PubMed Central

    You, Shingchern D.; Chen, Wei-Hwa; Chen, Woei-Kae

    2013-01-01

    This paper describes a multiresolution system based on MPEG-7 audio signature descriptors for music identification. Such an identification system may be used to detect illegally copied music circulated over the Internet. In the proposed system, low-resolution descriptors are used to search likely candidates, and then full-resolution descriptors are used to identify the unknown (query) audio. With this arrangement, the proposed system achieves both high speed and high accuracy. To deal with the problem that a piece of query audio may not be inside the system's database, we suggest two different methods to find the decision threshold. Simulation results show that the proposed method II can achieve an accuracy of 99.4% for query inputs both inside and outside the database. Overall, it is highly possible to use the proposed system for copyright control. PMID:23533359

  3. Using MPEG-7 audio descriptors for music querying

    NASA Astrophysics Data System (ADS)

    Gruhne, M.; Dittmar, C.

    2006-08-01

    Due to the growing amount of digital audio an increasing need to automatically categorize music and to create self-controlled and suitable playlists has been emerged. A few approaches to this task relying on low-level features have been published so far. Unfortunately the results utilizing those technologies are not sufficient yet. This paper gives an introduction how to enhance the results with regard to the perceptual similarity using different high-level descriptors and a powerful interaction between the algorithm and the user to consider his preferences. A successful interaction between server and client requires a powerful standardized query language. This paper describes the tools of the MPEG-7 Audio standard in detail and gives examples of already established query languages. Furthermore the requirements of a multimedia query language are identified and its application is exemplified by an automatic audio creation system using a query language.

  4. Say What? The Role of Audio in Multimedia Video

    NASA Astrophysics Data System (ADS)

    Linder, C. A.; Holmes, R. M.

    2011-12-01

    Audio, including interviews, ambient sounds, and music, is a critical-yet often overlooked-part of an effective multimedia video. In February 2010, Linder joined scientists working on the Global Rivers Observatory Project for two weeks of intensive fieldwork in the Congo River watershed. The team's goal was to learn more about how climate change and deforestation are impacting the river system and coastal ocean. Using stills and video shot with a lightweight digital SLR outfit and audio recorded with a pocket-sized sound recorder, Linder documented the trials and triumphs of working in the heart of Africa. Using excerpts from the six-minute Congo multimedia video, this presentation will illustrate how to record and edit an engaging audio track. Topics include interview technique, collecting ambient sounds, choosing and using music, and editing it all together to educate and entertain the viewer.

  5. Audio signal recognition for speech, music, and environmental sounds

    NASA Astrophysics Data System (ADS)

    Ellis, Daniel P. W.

    2003-10-01

    Human listeners are very good at all kinds of sound detection and identification tasks, from understanding heavily accented speech to noticing a ringing phone underneath music playing at full blast. Efforts to duplicate these abilities on computer have been particularly intense in the area of speech recognition, and it is instructive to review which approaches have proved most powerful, and which major problems still remain. The features and models developed for speech have found applications in other audio recognition tasks, including musical signal analysis, and the problems of analyzing the general ``ambient'' audio that might be encountered by an auditorily endowed robot. This talk will briefly review statistical pattern recognition for audio signals, giving examples in several of these domains. Particular emphasis will be given to common aspects and lessons learned.

  6. Multi-channel spatialization systems for audio signals

    NASA Technical Reports Server (NTRS)

    Begault, Durand R. (inventor)

    1993-01-01

    Synthetic head related transfer functions (HRTF's) for imposing reprogrammable spatial cues to a plurality of audio input signals included, for example, in multiple narrow-band audio communications signals received simultaneously are generated and stored in interchangeable programmable read only memories (PROM's) which store both head related transfer function impulse response data and source positional information for a plurality of desired virtual source locations. The analog inputs of the audio signals are filtered and converted to digital signals from which synthetic head related transfer functions are generated in the form of linear phase finite impulse response filters. The outputs of the impulse response filters are subsequently reconverted to analog signals, filtered, mixed, and fed to a pair of headphones.

  7. The Mad Hatter's Cocktail Party: A Social Mobile Audio Space Supporting Multiple Simultaneous Conversations

    E-print Network

    Aoki, Paul M.

    pervasive, continuous, near-toll-quality audio communication services to mobile users. The second is socialThe Mad Hatter's Cocktail Party: A Social Mobile Audio Space Supporting Multiple Simultaneous Pittsburgh, PA 15213-3890 USA ABSTRACT This paper presents a mobile audio space intended for use by gelled

  8. Potential and Limitations of a Teleteaching Environment based on H.323 Audio -Visual Communication Systems1

    E-print Network

    Stavrakakis, Ioannis

    the development of multimedia collaboration systems that are capable of providing real-time and high quality audio-visual communication systems [1]. While the Mbone tools provide for a cost effective, medium quality audio visual1 Potential and Limitations of a Teleteaching Environment based on H.323 Audio - Visual

  9. SINUSOIDAL MODELING OF SPOT MICROPHONE SIGNALS BASED ON NOISE TRANSPLANTATION FOR MULTICHANNEL AUDIO CODING

    E-print Network

    Tsakalides, Panagiotis

    -known sinu- soidal plus noise model (SNM). Sinusoids cannot be used per se for high-quality audio modeling AUDIO CODING Christos Tzagkarakis, Athanasios Mouchtaris, and Panagiotis Tsakalides Department focuses on high-fidelity multichannel audio model- ing based on an enhanced adaptation of the well

  10. Non-Reference Audio Quality Assessment for Online Live Music Recordings

    E-print Network

    Wang, Hsin-Min

    Non-Reference Audio Quality Assessment for Online Live Music Recordings Zhonghua Li1 , Ju for Internet users and the most prominent platform for sharing live music. The audio quality of this huge, location, and record- ing device. However, most video search engines do not take audio quality

  11. Narrowband Perceptual Audio Coding: Enhancements for Speech Hossein Najaf-Zadeh and Peter Kabal

    E-print Network

    Kabal, Peter

    Introduction Over the past decade, research in audio coding has been concentrated on high quality compression.g., wireless communications). Therefore, mod- erate audio quality at bit rates below 16 kbit/s is appro- priateNarrowband Perceptual Audio Coding: Enhancements for Speech Hossein Najaf-Zadeh and Peter Kabal

  12. Packet Loss Concealment for Multichannel Audio Using the Multiband Source/Filter Model

    E-print Network

    Tsakalides, Panagiotis

    information will result in an audible degradation of quality in the audio signal. This problem has been examPacket Loss Concealment for Multichannel Audio Using the Multiband Source/Filter Model Kiki {karadsun, mouchtar, tsakalid}@ics.forth.gr Abstract We recently proposed a multichannel audio coding method

  13. Automatic audio and manual transcripts alignment, time-code transfer and selection of exact transcripts

    E-print Network

    ): the quality of the automatic tran- script depends on the recognizer. If audio-related texts are availableAutomatic audio and manual transcripts alignment, time-code transfer and selection of exact focuses on automatic processing of sibling resources of audio and written documents, such as available

  14. A Multicore Operating System with QoS Guarantees for Network Audio Applications

    E-print Network

    Kubiatowicz, John D.

    proto- cols with quality-of-service (QoS) guarantees (e.g., Audio Video Bridging (AVB) [19]) and lowPAPERS A Multicore Operating System with QoS Guarantees for Network Audio Applications JUAN A Parallel Computing Laboratory, UC Berkeley, CA, USA 2 Center for New Music and Audio Technologies, UC

  15. Real-Time Implementation for Digital Watermarking in Audio Signals Using Perceptual Masking

    E-print Network

    Bruck, Jehoshua (Shuki)

    in audio quality. This causes severe problems of copyright violations and emphasizes the need for copyright to insert into its own media a digital watermark in such a way, that the audio quality is not reducedReal-Time Implementation for Digital Watermarking in Audio Signals Using Perceptual Masking Tal

  16. Hindawi Publishing Corporation EURASIP Journal on Audio, Speech, and Music Processing

    E-print Network

    Tsakalides, Panagiotis

    signals with good audio quality (above 3.0 perceptual grade compared to the original). It is demonstratedHindawi Publishing Corporation EURASIP Journal on Audio, Speech, and Music Processing Volume 2008 A multiresolution source/filter model for coding of audio source signals (spot recordings) is proposed. Spot

  17. Perceptual Audio Coding using Adaptive Pre-and Post-Filters and Lossless Compression

    E-print Network

    Yu, Bin

    coder (PAC) shows that the new method achieves a comparable compression ratio and audio quality quality audio signal is 32 - 48 kHz with an accuracy of 16 - 24 bits per sample. In particular, a CDÃ?ciently high compression ratios while maintaining a good audio quality for many applications. One example

  18. Investigating the Impact of Audio Degradations on Users: Subjective vs. Objective Assessment Methods

    E-print Network

    Sasse, Angela

    conferencing (MMC) is increasing in popularity, but it is often questioned whether the quality of the audio user community, determining the levels of audio and video quality required for users to effectively are widely used to assess audio and video quality. Typically, a short section of material is played, after

  19. Design and Analysis of a Scalable Watermarking Scheme for the Scalable Audio Coder

    E-print Network

    Sun, Qibin

    , data expansion property and perceptual quality. Index Terms-- watermarking, scalable audio coding of the average users, it is not suitable for applications such as audio archiving, studio, high-quality streaming1 Design and Analysis of a Scalable Watermarking Scheme for the Scalable Audio Coder Zhi Li

  20. An Efficient Scheme for Lossy Realtime Audio Data Compression 3922 (O-3) Robert C. Maher

    E-print Network

    Maher, Robert C.

    which retain much of the perceptual quality of the original signal. Existing lossy audio dataAn Efficient Scheme for Lossy Realtime Audio Data Compression 3922 (O-3) Robert C. Maher University be obtained by sending request and remittance to the Audio Engineering Society, 60 East 42nd St., New York

  1. On Studying the Impact of the Internet Delays on Audio Transmission

    E-print Network

    Wang, Yongge

    ,ehab,hhamed,brewster}@cs.depaul.edu Abstract-- The quality of the audio in IP telephony is significantly influenced by various factors on understanding the contribution of these factors on audio quality, and their impact on adaptive transport to quality degradation than the human eye. Hence maintenance of quality is more important in audio

  2. On the Impact of Loss and Delay Variation on Internet Packet Audio Transmission

    E-print Network

    Wang, Yongge

    to audio quality, and their impact on adaptive transport mechanisms such as error and buffer control. We observations to assess the audio quality of PCM µ-Law and G.728 codecs under different loss and delay applications. In the absence of such a guaran- tee of quality, research in audio transmission quality has

  3. COMPUTATIONALLY EFFICIENT COMPRESSION OF AUDIO SIGNALS BY MEANS OF RIQ-DPCM

    E-print Network

    Maher, Robert C.

    for high quality applications can achieve approximately 6:1 compression of monophonic audio materialCOMPUTATIONALLY EFFICIENT COMPRESSION OF AUDIO SIGNALS BY MEANS OF RIQ-DPCM Robert C. Maher well-known for speech signals are not applicable to arbitrary audio signals such as music. The new DPCM

  4. Hearing You Loud and Clear: Student Perspectives of Audio Feedback in Higher Education

    ERIC Educational Resources Information Center

    Gould, Jill; Day, Pat

    2013-01-01

    The use of audio feedback for students in a full-time community nursing degree course is appraised. The aim of this mixed methods study was to examine student views on audio feedback for written assignments. Questionnaires and a focus group were used to capture student opinion of this pilot project. The majority of students valued audio feedback…

  5. A Collaborative Interface for Multimodal Ink and Audio Documents Amit Regmi and Stephen M. Watt

    E-print Network

    Watt, Stephen M.

    A Collaborative Interface for Multimodal Ink and Audio Documents Amit Regmi and Stephen M. Watt and to archive multi-party communication sessions that involve audio and digital ink on a shared canvas be used to represent, to transmit, to record and to synchronize ink and audio channels. We find Ink

  6. AUDIO AUTHENTICATION BASED ON DISTRIBUTED SOURCE CODING David Varodayan, Yao-Chung Lin and Bernd Girod

    E-print Network

    Girod, Bernd

    to audio authentication. Our approach has similarities with secure biometric authentication [5AUDIO AUTHENTICATION BASED ON DISTRIBUTED SOURCE CODING David Varodayan, Yao-Chung Lin and Bernd-chung.lin, bgirod}@stanford.edu ABSTRACT Audio authentication is important in content delivery via untrusted

  7. Audio Use in E-Learning: What, Why, When, and How?

    ERIC Educational Resources Information Center

    Calandra, Brendan; Barron, Ann E.; Thompson-Sellers, Ingrid

    2008-01-01

    Decisions related to the implementation of audio in e-learning are perplexing for many instructional designers, and deciphering theory and principles related to audio use can be difficult for practitioners. Yet, as bandwidth on the Internet increases, digital audio is becoming more common in online courses. This article provides a review of…

  8. Design and Usability Testing of an Audio Platform Game for Players with Visual Impairments

    ERIC Educational Resources Information Center

    Oren, Michael; Harding, Chris; Bonebright, Terri L.

    2008-01-01

    This article reports on the evaluation of a novel audio platform game that creates a spatial, interactive experience via audio cues. A pilot study with players with visual impairments, and usability testing comparing the visual and audio game versions using both sighted players and players with visual impairments, revealed that all the…

  9. Journal information 1 Spatial audio quality perception (part 2): a linear

    E-print Network

    Jackson, Philip JB

    PAPERS Journal information 1 Spatial audio quality perception (part 2): a linear regression model from a range of spatial audio processes (SAPs), can be used to build a regression model of perceived spatial audio quality in terms of previously developed spatially and timbrally relevant metrics

  10. Journal information 1 Spatial audio quality perception (part 1): impact of

    E-print Network

    Jackson, Philip JB

    PAPERS Journal information 1 Spatial audio quality perception (part 1): impact of commonly at Harman Becker Automotive Systems GmbH, Germany Spatial audio processes (SAPs) commonly encountered in consumer audio reproduction systems are known to generate a range of impairments to spatial quality. Two

  11. Guided by Voices: An Audio Augmented Reality System Kent Lyons, Maribeth Gandy, Thad Starner

    E-print Network

    Starner, Thad E.

    Guided by Voices: An Audio Augmented Reality System Kent Lyons, Maribeth Gandy, Thad Starner, lightweight audio-only augmented reality infrastructure. The system uses a simple wearable computer and a RF in creating audio-only augmented reality games and show how our location infrastructure is generalizable

  12. Responding Effectively to Composition Students: Comparing Student Perceptions of Written and Audio Feedback

    ERIC Educational Resources Information Center

    Bilbro, J.; Iluzada, C.; Clark, D. E.

    2013-01-01

    The authors compared student perceptions of audio and written feedback in order to assess what types of students may benefit from receiving audio feedback on their essays rather than written feedback. Many instructors previously have reported the advantages they see in audio feedback, but little quantitative research has been done on how the…

  13. Audio-Visual Speech Recognition Based on AAM Parameter and Phoneme Analysis

    E-print Network

    Takiguchi, Tetsuya

    feature, of course the audio feature, greatly influences the recognition rate in these processingAudio-Visual Speech Recognition Based on AAM Parameter and Phoneme Analysis of Visual Feature Yuto,takigu}@kobe-u.ac.jp Abstract. As one of the techniques for robust speech recognition un- der noisy environment, audio

  14. Development and Assessment of Web Courses That Use Streaming Audio and Video Technologies.

    ERIC Educational Resources Information Center

    Ingebritsen, Thomas S.; Flickinger, Kathleen

    Iowa State University, through a program called Project BIO (Biology Instructional Outreach), has been using RealAudio technology for about 2 years in college biology courses that are offered entirely via the World Wide Web. RealAudio is a type of streaming media technology that can be used to deliver audio content and a variety of other media…

  15. Three dimensional audio versus head down TCAS displays

    NASA Technical Reports Server (NTRS)

    Begault, Durand R.; Pittman, Marc T.

    1994-01-01

    The advantage of a head up auditory display was evaluated in an experiment designed to measure and compare the acquisition time for capturing visual targets under two conditions: Standard head down traffic collision avoidance system (TCAS) display, and three-dimensional (3-D) audio TCAS presentation. Ten commercial airline crews were tested under full mission simulation conditions at the NASA Ames Crew-Vehicle Systems Research Facility Advanced Concepts Flight Simulator. Scenario software generated targets corresponding to aircraft which activated a 3-D aural advisory or a TCAS advisory. Results showed a significant difference in target acquisition time between the two conditions, favoring the 3-D audio TCAS condition by 500 ms.

  16. The sawing machine is a machine SAWING MACHINES

    E-print Network

    Gellman, Andrew J.

    wheel and an idler wheel support and drive the blade. POWER HACKSAW MACHINES DESCRIPTION All power-524 Some machines feed by gravity, the saw frame having weights that can be shifted to give greater or less that contains the saw band idler wheel, the drive motor switch, the tension adjustment handwheel and mechanism

  17. Hybrid Quantum Cloning Machine

    E-print Network

    Satyabrata Adhikari; A. K. Pati; Indranil Chakrabarty; B. S. Choudhury

    2007-06-14

    In this work, we introduce a special kind of quantum cloning machine called Hybrid quantum cloning machine. The introduced Hybrid quantum cloning machine or transformation is nothing but a combination of pre-existing quantum cloning transformations. In this sense it creates its own identity in the field of quantum cloners. Hybrid quantum cloning machine can be of two types: (i) State dependent and (ii) State independent or Universal. We study here the above two types of Hybrid quantum cloning machines. Later we will show that the state dependent hybrid quantum-cloning machine can be applied on only four input states. We will also find in this paper another asymmetric universal quantum cloning machine constructed from the combination of optimal universal B-H quantum cloning machine and universal anti-cloning machine. The fidelities of the two outputs are different and their values lie in the neighborhood of ${5/6} $

  18. Applying Machine Learning and Audio Analysis Techniques to Insect Recognition in Intelligent Traps

    E-print Network

    Batista, Gustavo

    intimately related to our wellbeing, in both positive and negative ways. For example, insect pests consume pest insects. However, over 98% of sprayed insecticides and 95% of herbicides reach a destination other, dengue and filariasis have shown that the vectors' resistance to insecticides has steadily increased

  19. IEEE ICASSP2003 AUDIO INFORMATION ACCESS FROM MEETING ROOMS

    E-print Network

    Edinburgh, University of

    . Spoken document retrieval (SDR) is a suc- cessful example of speech-as-text approaches. Evaluation to that obtained using human-generated reference tran- scripts, with little or no dependence on word error rateIEEE ICASSP­2003 Hong Kong AUDIO INFORMATION ACCESS FROM MEETING ROOMS Steve Renals University

  20. IEEE ICASSP--2003 AUDIO INFORMATION ACCESS FROM MEETING ROOMS

    E-print Network

    Edinburgh, University of

    . Spoken document retrieval (SDR) is a suc­ cessful example of speech­as­text approaches. Evaluation to that obtained using human­generated reference tran­ scripts, with little or no dependence on word error rateIEEE ICASSP--2003 Hong Kong AUDIO INFORMATION ACCESS FROM MEETING ROOMS Steve Renals University

  1. Adding Audio Description: Does It Make a Difference?

    ERIC Educational Resources Information Center

    Schmeidler, Emilie; Kirchner, Corinne

    2001-01-01

    A study involving 111 adults with blindness examined the impact of watching television science programs with and without audio description. Results indicate respondents gained and retained more information from watching programs with description. They reported that the description makes the program more enjoyable, interesting, and informative.…

  2. Recognition of Instrument Timbres in Real Polytimbral Audio Recordings

    E-print Network

    Ras, Zbigniew W.

    Recognition of Instrument Timbres in Real Polytimbral Audio Recordings Elzbieta Kubera1,2 , Alicja instruments play, information about timbre may help to separate melodic lines for automatic transcription of music [15] (spatial information might also be used here). Automatic recognition of timbre, i

  3. Melody Transcription from Music Audio: Approaches and Evaluation

    E-print Network

    Ellis, Dan

    1 Melody Transcription from Music Audio: Approaches and Evaluation Graham E. Poliner1, Daniel P use it might be. We go on to describe the results of full-scale evaluations of melody transcription achieve around 70% correct transcription at the frame level, including distinguishing between the presence

  4. Developing a Framework for Effective Audio Feedback: A Case Study

    ERIC Educational Resources Information Center

    Hennessy, Claire; Forrester, Gillian

    2014-01-01

    The increase in the use of technology-enhanced learning in higher education has included a growing interest in new approaches to enhance the quality of feedback given to students. Audio feedback is one method that has become more popular, yet evaluating its role in feedback delivery is still an emerging area for research. This paper is based on a…

  5. Digital Audio Broadcasting in the Short Wave Bands

    NASA Technical Reports Server (NTRS)

    Vaisnys, Arvydas

    1998-01-01

    For many decades the Short Wae broadcasting service has used high power, double-sideband AM signals to reach audiences far and wide. While audio quality was usually not very high, inexpensive receivers could be used to tune into broadcasts fro distant countries.

  6. Exploratory Evaluation of Audio Email Technology in Formative Assessment Feedback

    ERIC Educational Resources Information Center

    Macgregor, George; Spiers, Alex; Taylor, Chris

    2011-01-01

    Formative assessment generates feedback on students' performance, thereby accelerating and improving student learning. Anecdotal evidence gathered by a number of evaluations has hypothesised that audio feedback may be capable of enhancing student learning more than other approaches. In this paper we report on the preliminary findings of a…

  7. Sounds Good: Using Digital Audio for Evaluation Feedback

    ERIC Educational Resources Information Center

    Rotheram, Bob

    2009-01-01

    Feedback on student work is problematic for faculty and students in British higher education. Evaluation feedback takes faculty much time to produce and students are often dissatisfied with its quantity, timing, and clarity. The Sounds Good project has been experimenting with the use of digital audio for feedback, aiming to save faculty time and…

  8. Harmonia: Audio-Visual Correlation for Synchronous Expression Seth Hunter

    E-print Network

    Harmonia: Audio-Visual Correlation for Synchronous Expression Seth Hunter MIT Media Laboratory 20 Ames St. Cambridge MA, 02130 hunters@mit.edu ABSTRACT Historically, synchronous movements in collective gatherings have been practiced in nearly every culture. We hypothesize that paired computer music instruments

  9. Identity Authentication based on Audio Visual Biometrics: A Survey

    E-print Network

    Hua, Kien A.

    Identity Authentication based on Audio Visual Biometrics: A Survey Kai Li Department of Computer Science University of Central Florida Orlando, Florida 32816 Email: kaili@eecs.ucf.edu Abstract--Biometric authentication is an emerging tech- nology that utilize biometric data for the purpose of person identification

  10. Audio-visual perception system for a humanoid robotic head.

    PubMed

    Viciana-Abad, Raquel; Marfil, Rebeca; Perez-Lorenzo, Jose M; Bandera, Juan P; Romero-Garces, Adrian; Reche-Lopez, Pedro

    2014-01-01

    One of the main issues within the field of social robotics is to endow robots with the ability to direct attention to people with whom they are interacting. Different approaches follow bio-inspired mechanisms, merging audio and visual cues to localize a person using multiple sensors. However, most of these fusion mechanisms have been used in fixed systems, such as those used in video-conference rooms, and thus, they may incur difficulties when constrained to the sensors with which a robot can be equipped. Besides, within the scope of interactive autonomous robots, there is a lack in terms of evaluating the benefits of audio-visual attention mechanisms, compared to only audio or visual approaches, in real scenarios. Most of the tests conducted have been within controlled environments, at short distances and/or with off-line performance measurements. With the goal of demonstrating the benefit of fusing sensory information with a Bayes inference for interactive robotics, this paper presents a system for localizing a person by processing visual and audio data. Moreover, the performance of this system is evaluated and compared via considering the technical limitations of unimodal systems. The experiments show the promise of the proposed approach for the proactive detection and tracking of speakers in a human-robot interactive framework. PMID:24878593

  11. STEREOPHONIC PERSONAL AUDIO REPRODUCTION USING PLANARITY CONTROL OPTIMIZATION

    E-print Network

    Jackson, Philip JB

    STEREOPHONIC PERSONAL AUDIO REPRODUCTION USING PLANARITY CONTROL OPTIMIZATION Philip Coleman levels and complex phase patterns within the target zone. Furthermore, multichannel reproduction systems in an arbitrarily-placed target zone. In this study, the application of planarity control for stereo reproduction

  12. Jessika Reissland Crossmodal Integration of Audio-Visual Information

    E-print Network

    Kallenrode, May-Britt

    Jessika Reissland Crossmodal Integration of Audio-Visual Information PICS Publications of the Institute of Cognitive Science Volume 13-2010 #12;ISSN: 1610-5389 Series title: PICS Publications photo- graphs of animals, were presented to one side of a screen, while the participants fixated a cross

  13. Audio-Visual Perception System for a Humanoid Robotic Head

    PubMed Central

    Viciana-Abad, Raquel; Marfil, Rebeca; Perez-Lorenzo, Jose M.; Bandera, Juan P.; Romero-Garces, Adrian; Reche-Lopez, Pedro

    2014-01-01

    One of the main issues within the field of social robotics is to endow robots with the ability to direct attention to people with whom they are interacting. Different approaches follow bio-inspired mechanisms, merging audio and visual cues to localize a person using multiple sensors. However, most of these fusion mechanisms have been used in fixed systems, such as those used in video-conference rooms, and thus, they may incur difficulties when constrained to the sensors with which a robot can be equipped. Besides, within the scope of interactive autonomous robots, there is a lack in terms of evaluating the benefits of audio-visual attention mechanisms, compared to only audio or visual approaches, in real scenarios. Most of the tests conducted have been within controlled environments, at short distances and/or with off-line performance measurements. With the goal of demonstrating the benefit of fusing sensory information with a Bayes inference for interactive robotics, this paper presents a system for localizing a person by processing visual and audio data. Moreover, the performance of this system is evaluated and compared via considering the technical limitations of unimodal systems. The experiments show the promise of the proposed approach for the proactive detection and tracking of speakers in a human-robot interactive framework. PMID:24878593

  14. Improving Personal Diaries Using Social Audio Features Michael Kuhn

    E-print Network

    Improving Personal Diaries Using Social Audio Features Michael Kuhn Computer Engineering and Networks Laboratory ETH Zurich, Switzerland kuhnmi@tik.ee.ethz.ch Roger Wattenhofer Computer Engineering and Networks Laboratory ETH Zurich, Switzerland wattenhofer@tik.ee.ethz.ch Samuel Welten Computer Engineering

  15. Learning from Animated Concept Maps with Concurrent Audio Narration

    ERIC Educational Resources Information Center

    Nesbit, John C.; Adesope, Olusola O.

    2011-01-01

    An animated concept map is a presentation of a network diagram in which nodes and links are sequentially added or modified. An experiment compared learning from animated concept maps and text by randomly assigning 133 undergraduates to study 1 of 4 narrated animations presenting semantically equivalent information accompanied by identical audio

  16. AUDIO-BASED GENDER IDENTIFICATION USING BOOTSTRAPPING George Tzanetakis

    E-print Network

    Tzanetakis, George

    AUDIO-BASED GENDER IDENTIFICATION USING BOOTSTRAPPING George Tzanetakis gtzan@cs.uvic.ca Department gender identification is used for video indexing and can improve speech recognition results by using gender- specific classifiers. Gender identification in large datasets is difficult because of the large

  17. Audio-Visual Atoms for Generic Video Concept Classification

    E-print Network

    Chang, Shih-Fu

    to the diverse video content as well as the challenging condition such as uneven lighting, clutter, occlusions discriminative audio-visual codebooks are constructed for concept detection. Experiments over Kodak's consumer.1 [Information Storage and Retrieval] Content Analysis and Indexing; H.4.m [Information Systems Applications

  18. Enactive Mandala: Audio-visualizing Brain Waves Tomohiro Tokunaga

    E-print Network

    Lyons, Michael J.

    and animated visual music. Trans- parent real-time audio-visual feedback of brainwave quali- ties supports, this has largely reduced technical and financial barriers to getting involved in research on brainwaves a constructive approach to understanding brainwave data in the context of musical/artistic expression. Real

  19. The Role of Audio Media in the Lives of Children.

    ERIC Educational Resources Information Center

    Christenson, Peter G.; Lindlof, Thomas R.

    Mass communication researchers have largely ignored the role of audio media and popular music in the lives of children, yet the available evidence shows that children do listen. Extant studies yield a consistent developmental portrait of childrens' listening frequency, but there is a notable lack of programatic research over the past decade, one…

  20. Coding Overcomplete Representations of Audio using the MCLT

    E-print Network

    Yoon, Byung-Jun

    Coding Overcomplete Representations of Audio using the MCLT ByungJun Yoon California Institute complex lapped transform (MCLT). In general, it is difficult to encode signals using overcomplete be significantly reduced for MCLT-based representations, without the need for iterative methods of sparsity

  1. centre for digital music Blind Audio Source Separation

    E-print Network

    Plumbley, Mark

    for digital music Blind Audio Source Separation EMMANUEL VINCENT, MARIA G. JAFARI, SAMER A. ABDALLAH, MARK D Separation Emmanuel Vincent, Maria G. Jafari, Samer A. Abdallah, Mark D. Plumbley and Mike E. Davies Centre Separation (BASS) is the problem of recovering each source signal from a given mixture signal. This report

  2. 47 CFR 87.483 - Audio visual warning systems.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... audio visual warning system (AVWS) is a radar-based obstacle avoidance system. AVWS activates... exist. The continuously operating radar calculates the location, direction and groundspeed of nearby... obstacle. (a) Radiodetermination (radar) frequencies. Frequencies authorized under § 87.475(b)(8) of...

  3. BATS: The Blind Audio Tactile Mapping System Peter Parente

    E-print Network

    Bishop, Gary

    BATS: The Blind Audio Tactile Mapping System Peter Parente Department of Computer Science University of North Carolina at Chapel Hill gb@cs.unc.edu ABSTRACT The BATS project focuses on helping. The BATS project focuses on the latter area of research by helping people without sight access and explore

  4. Integrated Spacesuit Audio System Enhances Speech Quality and Reduces Noise

    NASA Technical Reports Server (NTRS)

    Huang, Yiteng Arden; Chen, Jingdong; Chen, Shaoyan Sharyl

    2009-01-01

    A new approach has been proposed for increasing astronaut comfort and speech capture. Currently, the special design of a spacesuit forms an extreme acoustic environment making it difficult to capture clear speech without compromising comfort. The proposed Integrated Spacesuit Audio (ISA) system is to incorporate the microphones into the helmet and use software to extract voice signals from background noise.

  5. AUDIO-VISUAL ISOLATED DIGIT RECOGNITION FOR WHISPERED SPEECH

    E-print Network

    Busso, Carlos

    AUDIO-VISUAL ISOLATED DIGIT RECOGNITION FOR WHISPERED SPEECH Xing Fan, Carlos Busso, and John H.L. Hansen Center for Robust Speech Systems(CRSS) University of Texas at Dallas, Richardson, Texas 75083 xxf of whisper, there are considerable differences between neutral and whispered speech in the spectral structure

  6. Audio-Visual Prosody: Perception, Detection, and Synthesis of Prominence

    E-print Network

    Beskow, Jonas

    Audio-Visual Prosody: Perception, Detection, and Synthesis of Prominence Samer Al Moubayed, Jonas-verbal interaction between the visual and the acoustic modalities from production and perception perspectives a study examining the perception of the behavior of the talking heads when gestures are added at pitch

  7. Business Information Audio Cassettes: Their Care and Feeding

    ERIC Educational Resources Information Center

    Noble, Valerie

    1973-01-01

    Audio cassettes raise new types of problems for special librarians. Confronted with sparse information about the media, the author, through trial and error, eventually produced basic guidelines and criteria for bibliographic control, storage, circulation, and material sources. These are described and workable solutions are presented. (5…

  8. Streaming Audio and Video: New Challenges and Opportunities for Museums.

    ERIC Educational Resources Information Center

    Spadaccini, Jim

    Streaming audio and video present new challenges and opportunities for museums. Streaming media is easier to author and deliver to Internet audiences than ever before; digital video editing is commonplace now that the tools--computers, digital video cameras, and hard drives--are so affordable; the cost of serving video files across the Internet…

  9. Chemical News Via Audio Tapes: Chemical Industry News

    ERIC Educational Resources Information Center

    Hanford, W. E.; And Others

    1972-01-01

    Tape coverage of internal R&D news now has a broader scope with improved features. A new tape series covering external news of broad interest has been initiated. The use of tape in a Continuing Education Program is discussed as the future plans for expanding the audio tape program. (1 reference) (Author)

  10. 1886th ISCA Workshop on Speech Synthesis, Bonn, Germany, August 22-24, 2007 Text Processingfor Text-to-SpeechSystems in Indian Languages

    E-print Network

    Black, Alan W

    -to-SpeechSystems in Indian Languages Anand Arokia Raj 1 , Tanuja Sarkar 1 , Satish Chandra Pammi 1 , SanthoshYuvaraj 1Technology, Hyderabad,India. 2 Indian Instituteof Technology, Kanpur, India. 3 Language of building text-to-speechsys- tems in Indian languages. 1. Introduction The objective of a text

  11. The Effect of Embedded Text-to-Speech and Vocabulary eBook Scaffolds on the Comprehension of Students with Reading Disabilities

    ERIC Educational Resources Information Center

    Gonzalez, Michelle

    2014-01-01

    Limited research exists concerning the effect of interactive electronic texts or eBooks on the reading comprehension of students with reading disabilities. The purpose of this study was to determine if there was a significant difference in oral retelling and comprehension performance on multiple-choice questions when 17 students with reading…

  12. Prahallad et al. / J Zhejiang Univ SCI 2005 6A(11):1229-1234 1229 A text to speech interface for Universal Digital Library

    E-print Network

    Black, Alan W

    Mellon University, Pittsburgh, PA 15217, USA) (2 International Institute of Information Technology system into UDL portal and describes the development process of Hindi, Telugu and Tamil voices under Festvox framework using unit selection techniques. We demonstrate the quality of the Tamil and Telugu

  13. Iowa Virtual Literacy Protocol: A Pre-Experimental Design Using Kurzweil 3000 Text-to-Speech Software with Incarcerated Adult Learners

    ERIC Educational Resources Information Center

    McCulley, Yvette K.

    2012-01-01

    The problem: The increasingly competitive global economy demands literate, educated workers. Both men and women experience the effects of education on employment rates and income. Racial and ethnic minorities, English language learners, and especially those with prison records are most deeply affected by the economic consequences of dropping out…

  14. Low Complexity Decoding in Parametric Stereo Audio Coding Run-Yu Tong and Pao-Chi Chang

    E-print Network

    Chang, Pao-Chi

    quality. II. BINAURAL CUES OF SPATIAL AUDIO CODING The ability of human auditory system (HAS) can localizeLow Complexity Decoding in Parametric Stereo Audio Coding Scheme Run-Yu Tong and Pao-Chi Chang (PS) is an audio coding object of MPEG-4 HE-AAC v2 which utilized the Spatial Audio Coding (SAC

  15. PARALLEL MODEL COMBINATION AND WORD RECOGNITION IN SOCCER AUDIO Jack H. Longton and Philip J.B. Jackson

    E-print Network

    Jackson, Philip JB

    PARALLEL MODEL COMBINATION AND WORD RECOGNITION IN SOCCER AUDIO Jack H. Longton and Philip J.longton@surrey.ac.uk] ABSTRACT The audio scene from broadcast soccer can be used for identifying highlights from the game. Audio to recognition performance. Index Terms-- Audio indexing, soccer, HMM 1. INTRODUCTION With growing distribution

  16. Maintaining high-quality IP audio services in lossy IP network environments

    NASA Astrophysics Data System (ADS)

    Barton, Robert J., III; Chodura, Hartmut

    2000-07-01

    In this paper we present our research activities in the area of digital audio processing and transmission. Today's available teleconference audio solutions are lacking in flexibility, robustness and fidelity. There was a need for enhancing the quality of audio for IP-based applications to guarantee optimal services under varying conditions. Multiple tests and user evaluations have shown that a reliable audio communication toolkit is essential for any teleconference application. This paper summarizes our research activities and gives an overview of developed applications. In a first step the parameters, which influence the audio quality, were evaluated. All of these parameters have to be optimized in order to result into the best achievable quality. Therefore it was necessary to enhance existing schemes or develop new methods. Applications were developed for Internet-Telephony, broadcast of live music and spatial audio for Virtual Reality environments. This paper describes these applications and issues of delivering high quality digital audio services over lossy IP networks.

  17. The method of narrow-band audio classification based on universal noise background model

    NASA Astrophysics Data System (ADS)

    Rui, Rui; Bao, Chang-chun

    2013-03-01

    Audio classification is the basis of content-based audio analysis and retrieval. The conventional classification methods mainly depend on feature extraction of audio clip, which certainly increase the time requirement for classification. An approach for classifying the narrow-band audio stream based on feature extraction of audio frame-level is presented in this paper. The audio signals are divided into speech, instrumental music, song with accompaniment and noise using the Gaussian mixture model (GMM). In order to satisfy the demand of actual environment changing, a universal noise background model (UNBM) for white noise, street noise, factory noise and car interior noise is built. In addition, three feature schemes are considered to optimize feature selection. The experimental results show that the proposed algorithm achieves a high accuracy for audio classification, especially under each noise background we used and keep the classification time less than one second.

  18. Hard Machinable Machining of Cobalt Super Alloys

    NASA Astrophysics Data System (ADS)

    ?ep, Robert; Janásek, Adam; Petr?, Jana; ?epová, Lenka; Sadílek, Marek; Kratochvíl, Ji?í

    2012-12-01

    The article deals with difficult-to-machine cobalt super alloys. The main aim is to test the basic properties of cobalt super alloys and propose suitable cutting materials and machining parameters under the designation 188 when machining. Although the development of technology in chipless machining such as moulding, precision casting and other manufacturing methods continues to advance, machining is still the leading choice for piece production, typical for energy and chemical engineering. Nowadays, super alloys are commonly used in turbine engines in regions that are subject to high temperatures, which require high strength, high temperature resistance, phase stability, as well as corrosion or oxidation resistance.

  19. Stirling machine operating experience

    NASA Technical Reports Server (NTRS)

    Ross, Brad; Dudenhoefer, James E.

    1991-01-01

    Numerous Stirling machines have been built and operated, but the operating experience of these machines is not well known. It is important to examine this operating experience in detail, because it largely substantiates the claim that Stirling machines are capable of reliable and lengthy lives. The amount of data that exists is impressive, considering that many of the machines that have been built are developmental machines intended to show proof of concept, and were not expected to operate for any lengthy period of time. Some Stirling machines (typically free-piston machines) achieve long life through non-contact bearings, while other Stirling machines (typically kinematic) have achieved long operating lives through regular seal and bearing replacements. In addition to engine and system testing, life testing of critical components is also considered.

  20. Deutsch Durch Audio-Visuelle Methode: An Audio-Lingual-Oral Approach to the Teaching of German.

    ERIC Educational Resources Information Center

    Dickinson Public Schools, ND. Instructional Media Center.

    This teaching guide, designed to accompany Chilton's "Deutsch Durch Audio-Visuelle Methode" for German 1 and 2 in a three-year secondary school program, focuses major attention on the operational plan of the program and a student orientation unit. A section on teaching a unit discusses four phases: (1) presentation, (2) explanation, (3)…

  1. IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, VOL. 10, NO. 5, JULY 2002 293 Musical Genre Classification of Audio Signals

    E-print Network

    Tzanetakis, George

    annotation is performed manually. Automatic musical genre classification can assist or replace the human user. This result is comparable to results reported for human musical genre classification. Index Terms Classification of Audio Signals George Tzanetakis, Student Member, IEEE, and Perry Cook, Member, IEEE Abstract

  2. Planning Schools for Use of Audio-Visual Materials. No. 3: The Audio-Visual Materials Center.

    ERIC Educational Resources Information Center

    National Education Association, Washington, DC. Dept. of Audiovisual Instruction.

    This manual discusses the role, organizational patterns, expected services, and space and housing needs of the audio-visual instructional materials center. In considering the housing of basic functions, photographs, floor layouts, diagrams, and specifications of equipment are presented. An appendix includes a 77-item bibliography, a 7-page list of…

  3. MACHINE PERCEPTION CZECH TECHNICAL

    E-print Network

    Kohli, Pushmeet

    CENTER FOR MACHINE PERCEPTION CZECH TECHNICAL UNIVERSITY RESEARCHREPORT ISSN1213-2365 Curvature for Machine Perception, K13133 FEE Czech Technical University 2 Microsoft Research Cambridge CTU­CMP­2011, Czech Technical University in Prague, No. 11, 2011 Published by Center for Machine Perception

  4. Talking Vending Machines

    E-print Network

    Hacker, Randi

    2011-02-02

    're in the central Tokyo Station and you're thirsty so you go to the, you know, vending machine to buy a drink. Only this is not your average vending machine. Uh uh. This vending machine has opinions on what beverage you want. Sensors built into large touch panels...

  5. DRILLING MACHINES GENERAL INFORMATION

    E-print Network

    Gellman, Andrew J.

    TC 9-524 Chapter 4 DRILLING MACHINES GENERAL INFORMATION PURPOSE This chapter contains basic information pertaining to drilling machines. A drilling machine comes in many shapes and sizes, from small hand-held power drills to bench mounted and finally floor-mounted models. They can perform operations

  6. Apprentice Machine Theory Outline.

    ERIC Educational Resources Information Center

    Connecticut State Dept. of Education, Hartford. Div. of Vocational-Technical Schools.

    This volume contains outlines for 16 courses in machine theory that are designed for machine tool apprentices. Addressed in the individual course outlines are the following topics: basic concepts; lathes; milling machines; drills, saws, and shapers; heat treatment and metallurgy; grinders; quality control; hydraulics and pneumatics;…

  7. Stochastic Optimization for Machine Learning

    E-print Network

    Powell, Warren B.

    Stochastic Optimization for Machine Learning ICML 2010, Haifa, Israel Tutorial by Nati Srebro Descent: formulation, analysis and use in machine learning · Learn about extensions and generalizations, and their Machine Learning counterparts Main Goal: Machine Learning is Stochastic Optimization #12;Outline

  8. IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, SPECIAL ISSUE ON DATA MINING OF SPEECH, AUDIO AND DIALOG 1 Mining Customer Care Dialogs for "Daily News"

    E-print Network

    Volinsky, Chris

    IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, SPECIAL ISSUE ON DATA MINING OF SPEECH, AUDIO describes the "VoiceTone Daily News" data mining tool for analyzing this information and presenting turns. Index Terms-- speech mining, data mining, spoken dialog systems, dialog success, business

  9. IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, SPECIAL ISSUE ON DATA MINING OF SPEECH, AUDIO AND DIALOG 1 Mining Customer Care Dialogs for ``Daily News''

    E-print Network

    Volinsky, Chris

    IEEE TRANSACTIONS ON SPEECH AND AUDIO PROCESSING, SPECIAL ISSUE ON DATA MINING OF SPEECH, AUDIO describes the ``VoiceTone Daily News'' data mining tool for analyzing this information and presenting turns. Index Terms--- speech mining, data mining, spoken dialog systems, dialog success, business

  10. Synthesis of audio spectra using a novel Bessel expansion

    NASA Astrophysics Data System (ADS)

    Vijayakumar, V.; Eswaran, C.

    2001-05-01

    It is shown in this paper that the intensity variations of an audio signal in the frequency domain can be expressed by using a novel mathematical function containing a series of weighted complex Bessel functions. By proper choice of values for two parameters, this function can transform an input spectrum of discrete frequencies of unit intensity into the known spectra of different musical instruments. Two specific examples of musical instruments are considered for evaluating the performance of this method. It is found that this function yields musical spectra with a good degree of accuracy. The proposed method is compared with known synthesis techniques such as FM, AFM, and DFM [J. M. Chowning, J. Audio Eng. Soc. 21, 526-534 (1973); Palamin et al., ibid. 36, 671 (1988); Tan, ibid. 42 (11), 918-926 (1994)]. A brief discussion on the physical basis for the derivation of the proposed function is also presented.

  11. Guidelines for the integration of audio cues into computer user interfaces

    SciTech Connect

    Sumikawa, D.A.

    1985-06-01

    Throughout the history of computers, vision has been the main channel through which information is conveyed to the computer user. As the complexities of man-machine interactions increase, more and more information must be transferred from the computer to the user and then successfully interpreted by the user. A logical next step in the evolution of the computer-user interface is the incorporation of sound and thereby using the sense of ''hearing'' in the computer experience. This allows our visual and auditory capabilities to work naturally together in unison leading to more effective and efficient interpretation of all information received by the user from the computer. This thesis presents an initial set of guidelines to assist interface developers in designing an effective sight and sound user interface. This study is a synthesis of various aspects of sound, human communication, computer-user interfaces, and psychoacoustics. We introduce the notion of an earcon. Earcons are audio cues used in the computer-user interface to provide information and feedback to the user about some computer object, operation, or interaction. A possible construction technique for earcons, the use of earcons in the interface, how earcons are learned and remembered, and the affects of earcons on their users are investigated. This study takes the point of view that earcons are a language and human/computer communication issue and are therefore analyzed according to the three dimensions of linguistics; syntactics, semantics, and pragmatics.

  12. Temporal Structure and Complexity Affect Audio-Visual Correspondence Detection

    PubMed Central

    Denison, Rachel N.; Driver, Jon; Ruff, Christian C.

    2013-01-01

    Synchrony between events in different senses has long been considered the critical temporal cue for multisensory integration. Here, using rapid streams of auditory and visual events, we demonstrate how humans can use temporal structure (rather than mere temporal coincidence) to detect multisensory relatedness. We find psychophysically that participants can detect matching auditory and visual streams via shared temporal structure for crossmodal lags of up to 200?ms. Performance on this task reproduced features of past findings based on explicit timing judgments but did not show any special advantage for perfectly synchronous streams. Importantly, the complexity of temporal patterns influences sensitivity to correspondence. Stochastic, irregular streams – with richer temporal pattern information – led to higher audio-visual matching sensitivity than predictable, rhythmic streams. Our results reveal that temporal structure and its complexity are key determinants for human detection of audio-visual correspondence. The distinctive emphasis of our new paradigms on temporal patterning could be useful for studying special populations with suspected abnormalities in audio-visual temporal perception and multisensory integration. PMID:23346067

  13. AudioSense: Enabling Real-time Evaluation of Hearing Aid Technology In-Situ.

    PubMed

    Hasan, Syed Shabih; Lai, Farley; Chipara, Octav; Wu, Yu-Hsiang

    2013-01-01

    AudioSense integrates mobile phones and web technology to measure hearing aid performance in real-time and in-situ. Measuring the performance of hearing aids in the real world poses significant challenges as it depends on the patient's listening context. AudioSense uses Ecological Momentary Assessment methods to evaluate both the perceived hearing aid performance as well as to characterize the listening environment using electronic surveys. AudioSense further characterizes a patient's listening context by recording their GPS location and sound samples. By creating a time-synchronized record of listening performance and listening contexts, AudioSense will allow researchers to understand the relationship between listening context and hearing aid performance. Performance evaluation shows that AudioSense is reliable, energy-efficient, and can estimate Signal-to-Noise Ratio (SNR) levels from captured audio samples. PMID:25013874

  14. AudioSense: Enabling Real-time Evaluation of Hearing Aid Technology In-Situ

    PubMed Central

    Hasan, Syed Shabih; Lai, Farley; Chipara, Octav; Wu, Yu-Hsiang

    2014-01-01

    AudioSense integrates mobile phones and web technology to measure hearing aid performance in real-time and in-situ. Measuring the performance of hearing aids in the real world poses significant challenges as it depends on the patient's listening context. AudioSense uses Ecological Momentary Assessment methods to evaluate both the perceived hearing aid performance as well as to characterize the listening environment using electronic surveys. AudioSense further characterizes a patient's listening context by recording their GPS location and sound samples. By creating a time-synchronized record of listening performance and listening contexts, AudioSense will allow researchers to understand the relationship between listening context and hearing aid performance. Performance evaluation shows that AudioSense is reliable, energy-efficient, and can estimate Signal-to-Noise Ratio (SNR) levels from captured audio samples. PMID:25013874

  15. Fault Tolerant State Machines

    NASA Technical Reports Server (NTRS)

    Burke, Gary R.; Taft, Stephanie

    2004-01-01

    State machines are commonly used to control sequential logic in FPGAs and ASKS. An errant state machine can cause considerable damage to the device it is controlling. For example in space applications, the FPGA might be controlling Pyros, which when fired at the wrong time will cause a mission failure. Even a well designed state machine can be subject to random errors us a result of SEUs from the radiation environment in space. There are various ways to encode the states of a state machine, and the type of encoding makes a large difference in the susceptibility of the state machine to radiation. In this paper we compare 4 methods of state machine encoding and find which method gives the best fault tolerance, as well as determining the resources needed for each method.

  16. Machine tool locator

    DOEpatents

    Hanlon, John A. (Los Alamos, NM); Gill, Timothy J. (Stanley, NM)

    2001-01-01

    Machine tools can be accurately measured and positioned on manufacturing machines within very small tolerances by use of an autocollimator on a 3-axis mount on a manufacturing machine and positioned so as to focus on a reference tooling ball or a machine tool, a digital camera connected to the viewing end of the autocollimator, and a marker and measure generator for receiving digital images from the camera, then displaying or measuring distances between the projection reticle and the reference reticle on the monitoring screen, and relating the distances to the actual position of the autocollimator relative to the reference tooling ball. The images and measurements are used to set the position of the machine tool and to measure the size and shape of the machine tool tip, and examine cutting edge wear. patent

  17. Omnidirectional Audio-Visual Talker Localization Based on Dynamic Fusion of Audio-Visual Features Using Validity and Reliability Criteria

    NASA Astrophysics Data System (ADS)

    Denda, Yuki; Nishiura, Takanobu; Yamashita, Yoichi

    This paper proposes a robust omnidirectional audio-visual (AV) talker localizer for AV applications. The proposed localizer consists of two innovations. One of them is robust omnidirectional audio and visual features. The direction of arrival (DOA) estimation using an equilateral triangular microphone array, and human position estimation using an omnidirectional video camera extract the AV features. The other is a dynamic fusion of the AV features. The validity criterion, called the audioor visual-localization counter, validates each audio- or visual-feature. The reliability criterion, called the speech arriving evaluator, acts as a dynamic weight to eliminate any prior statistical properties from its fusion procedure. The proposed localizer can compatibly achieve talker localization in a speech activity and user localization in a non-speech activity under the identical fusion rule. Talker localization experiments were conducted in an actual room to evaluate the effectiveness of the proposed localizer. The results confirmed that the talker localization performance of the proposed AV localizer using the validity and reliability criteria is superior to that of conventional localizers.

  18. Quantum Learning Machine

    E-print Network

    Jeongho Bang; James Lim; M. S. Kim; Jinhyoung Lee

    2008-03-31

    We propose a novel notion of a quantum learning machine for automatically controlling quantum coherence and for developing quantum algorithms. A quantum learning machine can be trained to learn a certain task with no a priori knowledge on its algorithm. As an example, it is demonstrated that the quantum learning machine learns Deutsch's task and finds itself a quantum algorithm, that is different from but equivalent to the original one.

  19. ASTP video tape recorder ground support equipment (audio/CTE splitter/interleaver). Operations manual

    NASA Technical Reports Server (NTRS)

    1974-01-01

    A descriptive handbook for the audio/CTE splitter/interleaver (RCA part No. 8673734-502) was presented. This unit is designed to perform two major functions: extract audio and time data from an interleaved video/audio signal (splitter section), and provide a test interleaved video/audio/CTE signal for the system (interleaver section). It is a rack mounting unit 7 inches high, 19 inches wide, 20 inches deep, mounted on slides for retracting from the rack, and weighs approximately 40 pounds. The following information is provided: installation, operation, principles of operation, maintenance, schematics and parts lists.

  20. Realization of guitar audio effects using methods of digital signal processing

    NASA Astrophysics Data System (ADS)

    Bu?, Szymon; Jedrzejewski, Konrad

    2015-09-01

    The paper is devoted to studies on possibilities of realization of guitar audio effects by means of methods of digital signal processing. As a result of research, some selected audio effects corresponding to the specifics of guitar sound were realized as the real-time system called Digital Guitar Multi-effect. Before implementation in the system, the selected effects were investigated using the dedicated application with a graphical user interface created in Matlab environment. In the second stage, the real-time system based on a microcontroller and an audio codec was designed and realized. The system is designed to perform audio effects on the output signal of an electric guitar.

  1. A Virtual Audio Guidance and Alert System for Commercial Aircraft Operations

    NASA Technical Reports Server (NTRS)

    Begault, Durand R.; Wenzel, Elizabeth M.; Shrum, Richard; Miller, Joel; Null, Cynthia H. (Technical Monitor)

    1996-01-01

    Our work in virtual reality systems at NASA Ames Research Center includes the area of aurally-guided visual search, using specially-designed audio cues and spatial audio processing (also known as virtual or "3-D audio") techniques (Begault, 1994). Previous studies at Ames had revealed that use of 3-D audio for Traffic Collision Avoidance System (TCAS) advisories significantly reduced head-down time, compared to a head-down map display (0.5 sec advantage) or no display at all (2.2 sec advantage) (Begault, 1993, 1995; Begault & Pittman, 1994; see Wenzel, 1994, for an audio demo). Since the crew must keep their head up and looking out the window as much as possible when taxiing under low-visibility conditions, and the potential for "blunder" is increased under such conditions, it was sensible to evaluate the audio spatial cueing for a prototype audio ground collision avoidance warning (GCAW) system, and a 3-D audio guidance system. Results were favorable for GCAW, but not for the audio guidance system.

  2. Debugging the virtual machine

    SciTech Connect

    Miller, P.; Pizzi, R.

    1994-09-02

    A computer program is really nothing more than a virtual machine built to perform a task. The program`s source code expresses abstract constructs using low level language features. When a virtual machine breaks, it can be very difficult to debug because typical debuggers provide only low level machine implementation in formation to the software engineer. We believe that the debugging task can be simplified by introducing aspects of the abstract design into the source code. We introduce OODIE, an object-oriented language extension that allows programmers to specify a virtual debugging environment which includes the design and abstract data types of the virtual machine.

  3. Chaotic Boltzmann machines

    PubMed Central

    Suzuki, Hideyuki; Imura, Jun-ichi; Horio, Yoshihiko; Aihara, Kazuyuki

    2013-01-01

    The chaotic Boltzmann machine proposed in this paper is a chaotic pseudo-billiard system that works as a Boltzmann machine. Chaotic Boltzmann machines are shown numerically to have computing abilities comparable to conventional (stochastic) Boltzmann machines. Since no randomness is required, efficient hardware implementation is expected. Moreover, the ferromagnetic phase transition of the Ising model is shown to be characterised by the largest Lyapunov exponent of the proposed system. In general, a method to relate probabilistic models to nonlinear dynamics by derandomising Gibbs sampling is presented. PMID:23558425

  4. Perspex machine II: visualization

    NASA Astrophysics Data System (ADS)

    Anderson, James A. D. W.

    2005-01-01

    We review the perspex machine and improve it by reducing its halting conditions to one condition. We also introduce a data structure, called the "access column," that can accelerate a wide class of perspex programs. We show how the perspex can be visualised as a tetrahedron, artificial neuron, computer program, and as a geometrical transformation. We discuss the temporal properties of the perspex machine, dissolve the famous time travel paradox, and present a hypothetical time machine. Finally, we discuss some mental properties and show how the perspex machine solves the mind-body problem and, specifically, how it provides one physical explanation for the occurrence of paradigm shifts.

  5. Perspex machine II: visualization

    NASA Astrophysics Data System (ADS)

    Anderson, James A. D. W.

    2004-12-01

    We review the perspex machine and improve it by reducing its halting conditions to one condition. We also introduce a data structure, called the "access column," that can accelerate a wide class of perspex programs. We show how the perspex can be visualised as a tetrahedron, artificial neuron, computer program, and as a geometrical transformation. We discuss the temporal properties of the perspex machine, dissolve the famous time travel paradox, and present a hypothetical time machine. Finally, we discuss some mental properties and show how the perspex machine solves the mind-body problem and, specifically, how it provides one physical explanation for the occurrence of paradigm shifts.

  6. Drum cutter mining machine

    SciTech Connect

    Oberste-beulmann, K.; Schupphaus, H.

    1980-02-19

    A drum cutter mining machine includes a machine frame with a winch having a drive wheel to engage a rack or chain which extends along the path of travel by the mining machine to propel the machine along a mine face. The mining machine is made up of discrete units which include a machine body and machine housings joined to opposite sides of the machine body. The winch is either coupled through a drive train with a feed drive motor or coupled to the drive motor for cutter drums. The machine housings each support a pivot shaft coupled by an arm to a drum cutter. One of these housings includes a removable end cover and a recess adapted to receive a support housing for a spur gear system used to transmit torque from a feed drive motor to a reduction gear system which is, in turn, coupled to the drive wheel of the winch. In one embodiment, a removable end cover on the machine housing provides access to the feed drive motor. The feed drive motor is arranged so that the rotational axis of its drive output shaft extends transversely to the stow side of the machine frame. In another embodiment, the reduction gear system is arranged at one side of the pivot shaft for the cutter drum while the drive motor therefor is arranged at the other side of the pivot shaft and coupled thereto through the spur gear system. In a further embodiment, the reduction gear system is disposed between the feed motor and the pivot shaft.

  7. BRITISH MOLDING MACHINE, PBQ AUTOMATIC COPE AND DRAG MOLDING MACHINE ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    BRITISH MOLDING MACHINE, PBQ AUTOMATIC COPE AND DRAG MOLDING MACHINE MAKES BOTH MOLD HALVES INDIVIDUALLY WHICH ARE LATER ROTATED, ASSEMBLED, AND LOWERED TO POURING CONVEYORS BY ASSISTING MACHINES. - Southern Ductile Casting Company, Casting, 2217 Carolina Avenue, Bessemer, Jefferson County, AL

  8. 14. Interior, Machine Shop, Roundhouse Machine Shop Extension, Southern Pacific ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    14. Interior, Machine Shop, Roundhouse Machine Shop Extension, Southern Pacific Railroad Carlin Shops, view to north (90mm lens). - Southern Pacific Railroad, Carlin Shops, Roundhouse Machine Shop Extension, Foot of Sixth Street, Carlin, Elko County, NV

  9. Diamond machine tool face lapping machine

    DOEpatents

    Yetter, H.H.

    1985-05-06

    An apparatus for shaping, sharpening and polishing diamond-tipped single-point machine tools. The isolation of a rotating grinding wheel from its driving apparatus using an air bearing and causing the tool to be shaped, polished or sharpened to be moved across the surface of the grinding wheel so that it does not remain at one radius for more than a single rotation of the grinding wheel has been found to readily result in machine tools of a quality which can only be obtained by the most tedious and costly processing procedures, and previously unattainable by simple lapping techniques.

  10. Machine Translation Project

    NASA Technical Reports Server (NTRS)

    Bajis, Katie

    1993-01-01

    The characteristics and capabilities of existing machine translation systems were examined and procurement recommendations were developed. Four systems, SYSTRAN, GLOBALINK, PC TRANSLATOR, and STYLUS, were determined to meet the NASA requirements for a machine translation system. Initially, four language pairs were selected for implementation. These are Russian-English, French-English, German-English, and Japanese-English.

  11. Friction-Testing Machine

    NASA Technical Reports Server (NTRS)

    Benz, F. J.; Dixon, D. S.; Shaw, R. C.

    1986-01-01

    Testing machine evaluates wear and ignition characteristics of materials in rubbing contact. Offers advantages over other laboratory methods of measuring wear because it simulates operating conditions under which material will actually be used. Machine used to determine wear characteristics, rank and select materials for service with such active oxidizers as oxygen, halogens, and oxides of nitrogen, measure wear characteristics, and determine coefficients of friction.

  12. Simple Machine Junk Cars

    ERIC Educational Resources Information Center

    Herald, Christine

    2010-01-01

    During the month of May, the author's eighth-grade physical science students study the six simple machines through hands-on activities, reading assignments, videos, and notes. At the end of the month, they can easily identify the six types of simple machine: inclined plane, wheel and axle, pulley, screw, wedge, and lever. To conclude this unit,…

  13. Simple Machines Made Simple.

    ERIC Educational Resources Information Center

    St. Andre, Ralph E.

    Simple machines have become a lost point of study in elementary schools as teachers continue to have more material to cover. This manual provides hands-on, cooperative learning activities for grades three through eight concerning the six simple machines: wheel and axle, inclined plane, screw, pulley, wedge, and lever. Most activities can be…

  14. Compound taper milling machine

    NASA Technical Reports Server (NTRS)

    Campbell, N. R.

    1969-01-01

    Simple, inexpensive milling machine tapers panels from a common apex to a uniform height at panel edge regardless of the panel perimeter configuration. The machine consists of an adjustable angled beam upon which the milling tool moves back and forth above a rotatable table upon which the workpiece is held.

  15. Technique for Machining Glass

    NASA Technical Reports Server (NTRS)

    Rice, S. H.

    1982-01-01

    Process for machining glass with conventional carbide tools requires a small quantity of a lubricant for aluminum applied to area of glass to be machined. A carbide tool is then placed against workpiece with light pressure. Tool is raised periodically to clear work of glass dust and particles. Additional lubricant is applied as it is displaced.

  16. Stirling machine operating experience

    SciTech Connect

    Ross, B.; Dudenhoefer, J.E.

    1994-09-01

    Numerous Stirling machines have been built and operated, but the operating experience of these machines is not well known. It is important to examine this operating experience in detail, because it largely substantiates the claim that stirling machines are capable of reliable and lengthy operating lives. The amount of data that exists is impressive, considering that many of the machines that have been built are developmental machines intended to show proof of concept, and are not expected to operate for lengthy periods of time. Some Stirling machines (typically free-piston machines) achieve long life through non-contact bearings, while other Stirling machines (typically kinematic) have achieved long operating lives through regular seal and bearing replacements. In addition to engine and system testing, life testing of critical components is also considered. The record in this paper is not complete, due to the reluctance of some organizations to release operational data and because several organizations were not contacted. The authors intend to repeat this assessment in three years, hoping for even greater participation.

  17. The Hooey Machine.

    ERIC Educational Resources Information Center

    Scarnati, James T.; Tice, Craig J.

    1992-01-01

    Describes how students can make and use Hooey Machines to learn how mechanical energy can be transferred from one object to another within a system. The Hooey Machine is made using a pencil, eight thumbtacks, one pushpin, tape, scissors, graph paper, and a plastic lid. (PR)

  18. Dictionary machine (for VLSI)

    SciTech Connect

    Ottmann, T.A.; Rosenberg, A.L.; Stockmeyer, L.J.

    1982-09-01

    The authors present the design of a dictionary machine that is suitable for VLSI implementation, and discusses how to realize this implementation efficiently. The machine supports the operations of search, insert, delete, and extractment on an arbitrary ordered set. Each of these operations takes time o(logn), where n is the number of entries present when the operation is performed. Moreover, arbitrary sequences of these instructions can be pipelined through the machine at a constant rate (i.e. independent of n and the capacity of the machine). The time o(logn) is an improvement over previous VLSI designs of dictionary machines which require time o(log n) per operation, where n is the maximum number of keys that can be stored. 10 references.

  19. 47 CFR 25.214 - Technical requirements for space stations in the Satellite Digital Audio Radio Service and...

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ...2014-10-01 false Technical requirements for space stations in the Satellite Digital Audio Radio Service and associated...Standards § 25.214 Technical requirements for space stations in the Satellite Digital Audio Radio Service and...

  20. On parallel machine scheduling 1

    E-print Network

    Magdeburg, Universität

    On parallel machine scheduling 1 machines with setup times. The setup has to be performed by a single server. The objective is to minimize even for the case of two identical parallel machines. This paper presents a pseudopolynomial

  1. Machine Learning ! ! ! ! !Srihari Neural Networks!

    E-print Network

    Machine Learning ! ! ! ! !Srihari 1 Neural Networks! #12;Machine Learning ! ! ! ! !Srihari 2 Two ! ! ! ! !Srihari 3 Neural Computation! Biological Motivation for Artificial Neural Networks #12;Machine Learning Groups Of Researchers in Neural Networks! 1. Study and model biological learning! · Network of neurons

  2. Informed spectral analysis: audio signal parameter estimation using side information

    NASA Astrophysics Data System (ADS)

    Fourer, Dominique; Marchand, Sylvain

    2013-12-01

    Parametric models are of great interest for representing and manipulating sounds. However, the quality of the resulting signals depends on the precision of the parameters. When the signals are available, these parameters can be estimated, but the presence of noise decreases the resulting precision of the estimation. Furthermore, the Cramér-Rao bound shows the minimal error reachable with the best estimator, which can be insufficient for demanding applications. These limitations can be overcome by using the coding approach which consists in directly transmitting the parameters with the best precision using the minimal bitrate. However, this approach does not take advantage of the information provided by the estimation from the signal and may require a larger bitrate and a loss of compatibility with existing file formats. The purpose of this article is to propose a compromised approach, called the 'informed approach,' which combines analysis with (coded) side information in order to increase the precision of parameter estimation using a lower bitrate than pure coding approaches, the audio signal being known. Thus, the analysis problem is presented in a coder/decoder configuration where the side information is computed and inaudibly embedded into the mixture signal at the coder. At the decoder, the extra information is extracted and is used to assist the analysis process. This study proposes applying this approach to audio spectral analysis using sinusoidal modeling which is a well-known model with practical applications and where theoretical bounds have been calculated. This work aims at uncovering new approaches for audio quality-based applications. It provides a solution for challenging problems like active listening of music, source separation, and realistic sound transformations.

  3. Music and audio - oh how they can stress your network

    NASA Astrophysics Data System (ADS)

    Fletcher, R.

    Nearly ten years ago a paper written by the Audio Engineering Society (AES)[1] made a number of interesting statements: 1. 2. The current Internet is inadequate for transmitting music and professional audio. Performance and collaboration across a distance stress beyond acceptable bounds the quality of service Audio and music provide test cases in which the bounds of the network are quickly reached and through which the defects in a network are readily perceived. Given these key points, where are we now? Have we started to solve any of the problems from the musician's point of view? What is it that musician would like to do that can cause the network so many problems? To understand this we need to appreciate that a trained musician's ears are extremely sensitive to very subtle shifts in temporal materials and localisation information. A shift of a few milliseconds can cause difficulties. So, can modern networks provide the temporal accuracy demanded at this level? The sample and bit rates needed to represent music in the digital domain is still contentious, but a general consensus in the professional world is for 96 KHz and IEEE 64-bit floating point. If this was to be run between two points on the network across 24 channels in near real time to allow for collaborative composition/production/performance, with QOS settings to allow as near to zero latency and jitter, it can be seen that the network indeed has to perform very well. Lighting the Blue Touchpaper for UK e-Science - Closing Conference of ESLEA Project The George Hotel, Edinburgh, UK 26-28 March, 200

  4. Machine tool evaluation and machining operation development

    SciTech Connect

    Morris, T.O.; Kegg, R.

    1997-03-15

    The purpose of this CRADA was to support Cincinnati Milacron`s needs in fabricating precision components, from difficult to machine materials, while maintaining and enhancing the precision manufacturing skills of the Oak Ridge Complex. Oak Ridge and Cincinnati Milacron personnel worked in a team relationship wherein each contributed equally to the success of the program. Process characterization, control technologies, machine tool capabilities, and environmental issues were the primary focus areas. In general, Oak Ridge contributed a wider range of expertise in machine tool testing and monitoring, and environmental testing on machining fluids to the defined tasks while Cincinnati Milacron personnel provided equipment, operations-specific knowledge and shop-floor services to each task. Cincinnati Milacron was very pleased with the results of all of the CRADA tasks. However, some of the environmental tasks were not carried through to a desired completion due to an expanding realization of need as the work progressed. This expansion of the desired goals then exceeded the time length of the CRADA. Discussions are underway on continuing these tasks under either a Work for Others agreement or some alternate funding.

  5. An assessment of individualized technical ear training for audio production.

    PubMed

    Kim, Sungyoung

    2015-07-01

    An individualized technical ear training method is compared to a non-individualized method. The efficacy of the individualized method is assessed using a standardized test conducted before and after the training period. Participants who received individualized training improved better than the control group on the test. Results indicate the importance of individualized training for acquisition of spectrum-identification and spectrum-matching skills. Individualized training, therefore, should be implemented by default into technical ear training programs used in audio production industry and education. PMID:26233051

  6. Audio-vocal responses elicited in adult cochlear implant users

    PubMed Central

    Loucks, Torrey M.; Suneel, Deepa; Aronoff, Justin M.

    2015-01-01

    Auditory deprivation experienced prior to receiving a cochlear implant could compromise neural connections that allow for modulation of vocalization using auditory feedback. In this report, pitch-shift stimuli were presented to adult cochlear implant users to test whether compensatory motor changes in vocal F0 could be elicited. In five of six participants, rapid adjustments in vocal F0 were detected following the stimuli, which resemble the cortically mediated pitch-shift responses observed in typical hearing individuals. These findings suggest that cochlear implants can convey vocal F0 shifts to the auditory pathway that might benefit audio-vocal monitoring. PMID:26520350

  7. Squeezing in the audio gravitational wave detection band

    E-print Network

    K. McKenzie; N. Grosse; W. P. Bowen; S. E. Whitcomb; M. B. Gray; D. E. McClelland; P. K. Lam

    2004-05-26

    We demonstrate the generation of broad-band continuous-wave optical squeezing down to 200Hz using a below threshold optical parametric oscillator (OPO). The squeezed state phase was controlled using a noise locking technique. We show that low frequency noise sources, such as seed noise, pump noise and detuning fluctuations, present in optical parametric amplifiers have negligible effect on squeezing produced by a below threshold OPO. This low frequency squeezing is ideal for improving the sensitivity of audio frequency measuring devices such as gravitational wave detectors.

  8. Fast Huffman encoding algorithms in MPEG-4 advanced audio coding

    NASA Astrophysics Data System (ADS)

    Brzuchalski, Grzegorz

    2014-11-01

    This paper addresses the optimisation problem of Huffman encoding in MPEG-4 Advanced Audio Coding stan- dard. At first, the Huffman encoding problem and the need of encoding two side info parameters scale factor and Huffman codebook are presented. Next, Two Loop Search, Maximum Noise Mask Ratio and Trellis Based algorithms of bit allocation are briefly described. Further, Huffman encoding optimisation are shown. New methods try to check and change scale factor bands as little as possible to estimate bitrate cost or its change. Finally, the complexity of old and new methods is calculated, compared and measured time of encoding is given.

  9. TV audio and video on the same channel

    NASA Technical Reports Server (NTRS)

    Hopkins, J. B.

    1979-01-01

    Transmitting technique adds audio to video signal during vertical blanking interval. SIVI (signal in the vertical interval) is used by TV networks and stations to transmit cuing and automatic-switching tone signals to augment automatic and manual operations. It can also be used to transmit one-way instructional information, such as bulletin alerts, program changes, and commercial-cutaway aural cues from the networks to affiliates. Additonally, it can be used as extra sound channel for second-language transmission to biligual stations.

  10. 2007 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics October 21-24, 2007, New Paltz, NY EXTENDING FINE-GRAIN SCALABLE AUDIO CODING TO VERY LOW BITRATES

    E-print Network

    Richard, Gaël

    . INTRODUCTION Depending on the target bitrate and/or quality, standard state-of- the-art audio coding (e.g. MPEG2007 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics October 21-24, 2007, New Paltz, NY EXTENDING FINE-GRAIN SCALABLE AUDIO CODING TO VERY LOW BITRATES USING OVERCOMPLETE

  11. 2005 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics October 16-19, 2005, New Paltz, NY A PROTOTYPE SYSTEM FOR OBJECT CODING OF MUSICAL AUDIO

    E-print Network

    Plumbley, Mark

    , such as MPEG4 AAC (Advanced Audio Coder), provide a near- transparent quality down to 50 Kb/s for mono signals2005 IEEE Workshop on Applications of Signal Processing to Audio and Acoustics October 16-19, 2005, New Paltz, NY A PROTOTYPE SYSTEM FOR OBJECT CODING OF MUSICAL AUDIO Emmanuel Vincent and Mark D

  12. Proc. of the 7th Int. Conference on Digital Audio Effects (DAFX-04), Naples, Italy, October 5-8, 2004 DIGITAL AUDIO EFFECTS APPLIED DIRECTLY ON A DSD BITSTREAM

    E-print Network

    Reiss, Josh

    high quality audio is often encoded in a one-bit, highly oversampled format, such as DSD. ProcessingProc. of the 7th Int. Conference on Digital Audio Effects (DAFX-04), Naples, Italy, October 5-8, 2004 DAFX-1 DIGITAL AUDIO EFFECTS APPLIED DIRECTLY ON A DSD BITSTREAM Josh Reiss and Mark Sandler

  13. 47 CFR 73.4275 - Tone clusters; audio attention-getting devices.

    Code of Federal Regulations, 2014 CFR

    2014-10-01

    ... 47 Telecommunication 4 2014-10-01 2014-10-01 false Tone clusters; audio attention-getting devices. 73.4275 Section 73.4275 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST... clusters; audio attention-getting devices. See Public Notice, FCC 76-610, dated July 2, 1976. 60 FCC 2d...

  14. 47 CFR 73.4275 - Tone clusters; audio attention-getting devices.

    Code of Federal Regulations, 2012 CFR

    2012-10-01

    ... 47 Telecommunication 4 2012-10-01 2012-10-01 false Tone clusters; audio attention-getting devices. 73.4275 Section 73.4275 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST... clusters; audio attention-getting devices. See Public Notice, FCC 76-610, dated July 2, 1976. 60 FCC 2d...

  15. 47 CFR 73.4275 - Tone clusters; audio attention-getting devices.

    Code of Federal Regulations, 2013 CFR

    2013-10-01

    ... 47 Telecommunication 4 2013-10-01 2013-10-01 false Tone clusters; audio attention-getting devices. 73.4275 Section 73.4275 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST... clusters; audio attention-getting devices. See Public Notice, FCC 76-610, dated July 2, 1976. 60 FCC 2d...

  16. 47 CFR 73.4275 - Tone clusters; audio attention-getting devices.

    Code of Federal Regulations, 2011 CFR

    2011-10-01

    ... 47 Telecommunication 4 2011-10-01 2011-10-01 false Tone clusters; audio attention-getting devices. 73.4275 Section 73.4275 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST... clusters; audio attention-getting devices. See Public Notice, FCC 76-610, dated July 2, 1976. 60 FCC 2d...

  17. TravelSampo System for Creating Mobile Audio Guide Tours Enriched with Linked Data

    E-print Network

    Hyvönen, Eero

    with traditional audio guide systems. For the museums, TravelSampo offers a cost-efficient and dynamic wayTravelSampo System for Creating Mobile Audio Guide Tours Enriched with Linked Data Eero Hyvönen, Aleksi Lindblad, and Eetu Mäkelä Aalto University, Helsinki TravelSampo [1] is a prototype system

  18. The SWRL Audio Laboratory System (ALS): An Integrated Configuration for Psychomusicology Research. Technical Report 51.

    ERIC Educational Resources Information Center

    Williams, David Brian; Hoskin, Richard K.

    This report describes features of the Audio Laboratory System (ALS), a device which supports research activities of the Southwest Regional Laboratory's Music Program. The ALS is used primarily to generate recorded audio tapes for psychomusicology research related to children's perception and learning of music concepts such as pitch, loudness,…

  19. Students' Attitudes to and Usage of Academic Feedback Provided via Audio Files

    ERIC Educational Resources Information Center

    Merry, Stephen; Orsmond, Paul

    2008-01-01

    This study explores students' attitudes to the provision of formative feedback on academic work using audio files together with the ways in which students implement such feedback within their learning. Fifteen students received audio file feedback on written work and were subsequently interviewed regarding their utilisation of that feedback within…

  20. Active Learning in the Online Environment: The Integration of Student-Generated Audio Files

    ERIC Educational Resources Information Center

    Bolliger, Doris U.; Armier, David Des, Jr.

    2013-01-01

    Educators have integrated instructor-produced audio files in a variety of settings and environments for purposes such as content presentation, lecture reviews, student feedback, and so forth. Few instructors, however, require students to produce audio files and share them with peers. The purpose of this study was to obtain empirical data on…

  1. "Listen to This!" Utilizing Audio Recordings to Improve Instructor Feedback on Writing in Mathematics

    ERIC Educational Resources Information Center

    Weld, Christopher

    2014-01-01

    Providing audio files in lieu of written remarks on graded assignments is arguably a more effective means of feedback, allowing students to better process and understand the critique and improve their future work. With emerging technologies and software, this audio feedback alternative to the traditional paradigm of providing written comments…

  2. 47 CFR Figure 2 to Subpart N of... - Typical Audio Wave

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 47 Telecommunication 1 2010-10-01 2010-10-01 false Typical Audio Wave 2 Figure 2 to Subpart N of Part 2 Telecommunication FEDERAL COMMUNICATIONS COMMISSION GENERAL FREQUENCY ALLOCATIONS AND RADIO... Audio Wave EC03JN91.006...

  3. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2013 CFR

    2013-04-01

    ... video material. 232.304 Section 232.304 Commodity and Securities Exchanges SECURITIES AND EXCHANGE... Submissions § 232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio or video material in a document delivered to investors and others that is not reproduced in...

  4. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2012 CFR

    2012-04-01

    ... video material. 232.304 Section 232.304 Commodity and Securities Exchanges SECURITIES AND EXCHANGE... Submissions § 232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio or video material in a document delivered to investors and others that is not reproduced in...

  5. 17 CFR 232.304 - Graphic, image, audio and video material.

    Code of Federal Regulations, 2014 CFR

    2014-04-01

    ... video material. 232.304 Section 232.304 Commodity and Securities Exchanges SECURITIES AND EXCHANGE... Submissions § 232.304 Graphic, image, audio and video material. (a) If a filer includes graphic, image, audio or video material in a document delivered to investors and others that is not reproduced in...

  6. A Fast Data Structure for Disk-Based Audio Editing Dominic Mazzoni and Roger B. Dannenberg

    E-print Network

    Dannenberg, Roger B.

    A Fast Data Structure for Disk-Based Audio Editing Dominic Mazzoni and Roger B. Dannenberg School or reversed later. However, these types of editors force users to keep all of the original audio clips at any time. This also means that the user can safely throw away the original files or modify them

  7. A METHOD FOR IDENTIFYING REPETITION STRUCTURE IN MUSICAL AUDIO BASED ON TIME SERIES PREDICTION

    E-print Network

    Dixon, Simon

    A METHOD FOR IDENTIFYING REPETITION STRUCTURE IN MUSICAL AUDIO BASED ON TIME SERIES PREDICTION annotation as a source of information. In particular, the task of music structure analysis has re- cently a frequently occurring audio segment [4]. In this work, we propose a novel approach for detecting repetition

  8. IMPROVING HANDS-FREE SPEECH RECOGNITION IN A CAR THROUGH AUDIO-VISUAL VOICE ACTIVITY DETECTION

    E-print Network

    IMPROVING HANDS-FREE SPEECH RECOGNITION IN A CAR THROUGH AUDIO-VISUAL VOICE ACTIVITY DETECTION the speech recognition perfor- mance in a noisy car environment can be improved by com- bining audio-visual voice activity detection (VAD) with mi- crophone array processing techniques. That is accomplished

  9. Guidelines for the Production of Audio Materials for Print Handicapped Readers.

    ERIC Educational Resources Information Center

    National Library of Australia, Canberra.

    Procedural guidelines developed by the Audio Standards Committee of the National Library of Australia to help improve the overall quality of production of audio materials for visually handicapped readers are presented. This report covers the following areas: selection of narrators and the narration itself; copyright; recording of books, magazines,…

  10. Effects of Audio-Visual Information on the Intelligibility of Alaryngeal Speech

    ERIC Educational Resources Information Center

    Evitts, Paul M.; Portugal, Lindsay; Van Dine, Ami; Holler, Aline

    2010-01-01

    Background: There is minimal research on the contribution of visual information on speech intelligibility for individuals with a laryngectomy (IWL). Aims: The purpose of this project was to determine the effects of mode of presentation (audio-only, audio-visual) on alaryngeal speech intelligibility. Method: Twenty-three naive listeners were…

  11. Rethinking the Red Ink: Audio-Feedback in the ESL Writing Classroom.

    ERIC Educational Resources Information Center

    Johanson, Robert

    1999-01-01

    This paper describes audio-feedback as a teaching method for English-as-a-Second-Language (ESL) writing classes. Using this method, writing instructors respond to students' compositions by recording their comments onto an audiocassette, then returning the paper and cassette to the students. The first section describes audio-feedback and explains…

  12. A Portable Audio/Video Recorder for Longitudinal Study of Child Development

    E-print Network

    Roy, Deb

    A Portable Audio/Video Recorder for Longitudinal Study of Child Development Soroush Vosoughi MIT- istic, longitudinal recordings of child development. Categories and Subject Descriptors H.5.2 [User, privacy man- agement, video recorder, audio recorder, child development 1. INTRODUCTION Collection

  13. LiveDescribe: Can Amateur Describers Create High-Quality Audio Description?

    ERIC Educational Resources Information Center

    Branje, Carmen J.; Fels, Deborah I.

    2012-01-01

    Introduction: The study presented here evaluated the usability of the audio description software LiveDescribe and explored the acceptance rates of audio description created by amateur describers who used LiveDescribe to facilitate the creation of their descriptions. Methods: Twelve amateur describers with little or no previous experience with…

  14. Social-and Interactive-Television Applications Based on Real-Time Ambient-Audio Identification

    E-print Network

    Cortes, Corinna

    Social- and Interactive-Television Applications Based on Real-Time Ambient-Audio Identification applications for mass personalization: personalized content layers, ad hoc social communities, real. Our audio identification system does not use dedicated interactive TV hardware and does not compromise

  15. 47 CFR 73.4275 - Tone clusters; audio attention-getting devices.

    Code of Federal Regulations, 2010 CFR

    2010-10-01

    ... 47 Telecommunication 4 2010-10-01 2010-10-01 false Tone clusters; audio attention-getting devices. 73.4275 Section 73.4275 Telecommunication FEDERAL COMMUNICATIONS COMMISSION (CONTINUED) BROADCAST... clusters; audio attention-getting devices. See Public Notice, FCC 76-610, dated July 2, 1976. 60 FCC 2d...

  16. 37 CFR 201.27 - Initial notice of distribution of digital audio recording devices or media.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... distribution of digital audio recording devices or media. 201.27 Section 201.27 Patents, Trademarks, and Copyrights COPYRIGHT OFFICE, LIBRARY OF CONGRESS COPYRIGHT OFFICE AND PROCEDURES GENERAL PROVISIONS § 201.27 Initial notice of distribution of digital audio recording devices or media. (a) General. This...

  17. 37 CFR 201.27 - Initial notice of distribution of digital audio recording devices or media.

    Code of Federal Regulations, 2011 CFR

    2011-07-01

    ... distribution of digital audio recording devices or media. 201.27 Section 201.27 Patents, Trademarks, and Copyrights COPYRIGHT OFFICE, LIBRARY OF CONGRESS COPYRIGHT OFFICE AND PROCEDURES GENERAL PROVISIONS § 201.27 Initial notice of distribution of digital audio recording devices or media. (a) General. This...

  18. 37 CFR 201.28 - Statements of Account for digital audio recording devices or media.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... 37 Patents, Trademarks, and Copyrights 1 2014-07-01 2014-07-01 false Statements of Account for digital audio recording devices or media. 201.28 Section 201.28 Patents, Trademarks, and Copyrights U.S. COPYRIGHT OFFICE, LIBRARY OF CONGRESS COPYRIGHT OFFICE AND PROCEDURES GENERAL PROVISIONS § 201.28 Statements of Account for digital audio...

  19. 37 CFR 201.28 - Statements of Account for digital audio recording devices or media.

    Code of Federal Regulations, 2013 CFR

    2013-07-01

    ... digital audio recording devices or media. 201.28 Section 201.28 Patents, Trademarks, and Copyrights COPYRIGHT OFFICE, LIBRARY OF CONGRESS COPYRIGHT OFFICE AND PROCEDURES GENERAL PROVISIONS § 201.28 Statements of Account for digital audio recording devices or media. (a) General. This section prescribes...

  20. 37 CFR 201.28 - Statements of Account for digital audio recording devices or media.

    Code of Federal Regulations, 2012 CFR

    2012-07-01

    ... 37 Patents, Trademarks, and Copyrights 1 2012-07-01 2012-07-01 false Statements of Account for digital audio recording devices or media. 201.28 Section 201.28 Patents, Trademarks, and Copyrights COPYRIGHT OFFICE, LIBRARY OF CONGRESS COPYRIGHT OFFICE AND PROCEDURES GENERAL PROVISIONS § 201.28 Statements of Account for digital audio...

  1. IEEE SIGNAL PROCESSING MAGAZINE [84] MARCH 2009 he field of audio forensics involves many topics

    E-print Network

    Maher, Robert C.

    IEEE SIGNAL PROCESSING MAGAZINE [84] MARCH 2009 T he field of audio forensics involves many topics by applying modern DSP theory to problems of interest to the forensics community, and this article is written in audio forensic laboratories. However, this article must also present several of the frustrations and pit

  2. SECTION-LEVEL MODELING OF MUSICAL AUDIO FOR LINKING PERFORMANCES TO SCORES IN TURKISH MAKAM MUSIC

    E-print Network

    SECTION-LEVEL MODELING OF MUSICAL AUDIO FOR LINKING PERFORMANCES TO SCORES IN TURKISH MAKAM MUSIC, for instance, in Turkish makam music. In addition to hav- ing low computational complexity compared to note method outperforms a previously presented approach on a Turkish makam music corpus. Index Terms-- Audio

  3. Mzik Trlerinin Co-MRMR ile Siniflandirilmasi Audio Genre Classification with Co-MRMR

    E-print Network

    Cataltepe, Zehra

    Müzik Türlerinin Co-MRMR ile Siniflandirilmasi Audio Genre Classification with Co-MRMR Yusuf Yaslan geleneksel Birlikte Örenme algoritmasindan daha iyi sonuç vermitir. Abstract In a classification problem classifiers trained on different feature subsets for Co-training is used for audio music genre classification

  4. AudioVisual Speaker Detection using Dynamic Bayesian Networks Submission no. 182

    E-print Network

    Pavlovic, Vladimir

    -the-shelf" visual and audio sensors (face, skin, texture, mouth motion, and silence detectors) are optimally fused of multiple sensors can be efficiently formulated using dynamic Bayesian net- works (DBNs). DBN framework and audio sensors, contextual information, temporal constraints and one's expert knowledge in solv- ing

  5. Audio--Visual Speaker Detection using Dynamic Bayesian Networks Submission no. 182

    E-print Network

    Pavlovic, Vladimir

    ­the­shelf'' visual and audio sensors (face, skin, texture, mouth motion, and silence detectors) are optimally fused of multiple sensors can be efficiently formulated using dynamic Bayesian net­ works (DBNs). DBN framework and audio sensors, contextual information, temporal constraints and one's expert knowledge in solv­ ing

  6. 36 CFR 5.5 - Commercial filming, still photography, and audio recording.

    Code of Federal Regulations, 2014 CFR

    2014-07-01

    ... schedule for still photography conducted under a permit issued under 43 CFR part 5 applies to audio... of 43 CFR part 5, subpart A. Failure to comply with any provision of 43 CFR part 5 is a violation of... photography, and audio recording. 5.5 Section 5.5 Parks, Forests, and Public Property NATIONAL PARK...

  7. AUTOMATIC LANGUAGE IDENTIFICATION IN MUSIC VIDEOS WITH LOW LEVEL AUDIO AND VISUAL FEATURES

    E-print Network

    Cortes, Corinna

    AUTOMATIC LANGUAGE IDENTIFICATION IN MUSIC VIDEOS WITH LOW LEVEL AUDIO AND VISUAL FEATURES Vijay LID in speech. Here, we study the problem of LID in music videos uploaded on YouTube. We use a "bag-of-words" approach based on state- of-the-art content based audio-visual features and linear SVM classifiers

  8. 37 CFR 201.27 - Initial notice of distribution of digital audio recording devices or media.

    Code of Federal Regulations, 2010 CFR

    2010-07-01

    ... integrated combination units (“boomboxes”); portable personal recorders; stand-alone home recorders (“tape... required by section 1003(b) of the Audio Home Recording Act of 1992, Public Law 102-563, title 17 of the... Notice is a notice under section 1003(b) of the Audio Home Recording Act of 1992, Public Law...

  9. Digital audio watermarking using Moment-Preserving Thresholding DooSeop Choia

    E-print Network

    Kim, Taejeong

    Digital audio watermarking using Moment-Preserving Thresholding DooSeop Choia , Hae Kyung Junga for digital images has been used in digital image processing for decades, especially in image binarization attacks including MP3 compression, re-sampling, jittering, and, DA/AD conversion. Keywords: Digital audio

  10. FINGERPRINTING TO IDENTIFY REPEATED SOUND EVENTS IN LONG-DURATION PERSONAL AUDIO RECORDINGS

    E-print Network

    Ellis, Dan

    FINGERPRINTING TO IDENTIFY REPEATED SOUND EVENTS IN LONG-DURATION PERSONAL AUDIO RECORDINGS James P for this data is a challenge; in this paper we investigate the use of an audio fingerprinting technique recurrent sound events within long (multi-day) recordings. The fingerprint- ing technique is based on energy

  11. Lossless Audio Compression: A Case Study CISE Technical Report 08-415

    E-print Network

    Mishra, Prabhat

    and increasing internet bandwidth. This paper investigates improvements to the Free Lossless Audio Codec (FLAC the blocking size in FLAC shows minor improvements in compression efficiency. Cristobal Rivero cristobalrivero popular. Through the widespread use of high-speed internet, consumers can download and transfer audio

  12. ORACLE ESTIMATION OF ADAPTIVE COSINE PACKET TRANSFORMS FOR UNDERDETERMINED AUDIO SOURCE SEPARATION

    E-print Network

    Plumbley, Mark

    ORACLE ESTIMATION OF ADAPTIVE COSINE PACKET TRANSFORMS FOR UNDERDETERMINED AUDIO SOURCE SEPARATION the problem of instantaneous, underdetermined audio source separation by time-frequency masking. Using oracle sources available, and that we know, or have estimated, the mixing structure. Oracle estimation of four

  13. 76 FR 591 - Determination of Rates and Terms for Preexisting Subscription and Satellite Digital Audio Radio...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-01-05

    ..., respectively. 72 FR 71795 (December 19, 2007), 73 FR 4080 (January 24, 2008). Section 804(b)(3)(B) of the... Royalty Board Determination of Rates and Terms for Preexisting Subscription and Satellite Digital Audio... subscription and satellite digital audio radio services for the digital performance of sound recordings and...

  14. Planning Schools for Use of Audio-Visual Materials. No. 1--Classrooms, 3rd Edition.

    ERIC Educational Resources Information Center

    National Education Association, Washington, DC.

    Intended to inform school board administrators and teachers of the current (1958) thinking on audio-visual instruction for use in planning new buildings, purchasing equipment, and planning instruction. Attention is given the problem of overcoming obstacles to the incorporation of audio-visual materials into the curriculum. Discussion includes--(1)…

  15. Energetic and informational masking effects in an audio-visual speech recognition system.

    E-print Network

    Barker, Jon

    1 Energetic and informational masking effects in an audio-visual speech recognition system. Jon Barker and Xu Shao Abstract The paper presents a robust audio-visual speech recognition technique called on a small vocabulary simultaneous speech recognition task in conditions that promote two contrasting types

  16. Integration of Text and Audio Features for Genre Classification in Music Information

    E-print Network

    Rauber,Andreas

    be a song's audio features as well as its lyrics. Both of these modalities have their advantages as text may, and browsing access by perceived sound similarity. Song lyrics cover semantic information about a song.e. automatically assigning musical genres to tracks based on audio features as well as content words in song lyrics

  17. CAUSAL TEMPO TRACKING OF AUDIO Matthew E. P. Davies and Mark D. Plumbley

    E-print Network

    Plumbley, Mark

    CAUSAL TEMPO TRACKING OF AUDIO Matthew E. P. Davies and Mark D. Plumbley Centre for Digital Music Queen Mary University of London ABSTRACT We introduce a causal approach to tempo tracking for mu- sical high- level knowledge of the musical audio. The tempo track- ing system is divided into two sections

  18. An Interactive Concert Program Based on Infrared Watermark and Audio Synthesis

    NASA Astrophysics Data System (ADS)

    Wang, Hsi-Chun; Lee, Wen-Pin Hope; Liang, Feng-Ju

    The objective of this research is to propose a video/audio system which allows the user to listen the typical music notes in the concert program under infrared detection. The system synthesizes audio with different pitches and tempi in accordance with the encoded data in a 2-D barcode embedded in the infrared watermark. The digital halftoning technique has been used to fabricate the infrared watermark composed of halftone dots by both amplitude modulation (AM) and frequency modulation (FM). The results show that this interactive system successfully recognizes the barcode and synthesizes audio under infrared detection of a concert program which is also valid for human observation of the contents. This interactive video/audio system has greatly expanded the capability of the printout paper to audio display and also has many potential value-added applications.

  19. 78 FR 44029 - Establishment of Rules and Policies for the Digital Audio Radio Satellite Service in the 2310...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2013-07-23

    ...Rules and Policies for the Digital Audio Radio Satellite Service in the 2310-2360 MHz...collections for Satellite Digital Audio Radio Service (SDARS) terrestrial repeaters...Rules and Policies for the Digital Audio Radio Satellite Service in the 2310-2360...

  20. Proc. of the 12th Int. Conference on Digital Audio Effects (DAFx-09), Como, Italy, September 1-4, 2009

    E-print Network

    Plumbley, Mark

    to produce a high quality representation of audio using similar techniques. Rather than defining newProc. of the 12th Int. Conference on Digital Audio Effects (DAFx-09), Como, Italy, September 1-4, 2009 ESTIMATING PARAMETERS FROM AUDIO FOR AN EG+LFO MODEL OF PITCH ENVELOPES Stephen J. Welburn

  1. Proc. of the 11th Int. Conference on Digital Audio Effects (DAFx-08), Espoo, Finland, September 1-4, 2008

    E-print Network

    Plumbley, Mark

    may become smeared or artificial, resulting in a perceptual loss of audio quality. As a secondaryProc. of the 11th Int. Conference on Digital Audio Effects (DAFx-08), Espoo, Finland, September 1-4, 2008 AUTOMATED RHYTHMIC TRANSFORMATION OF MUSICAL AUDIO Jason A. Hockman CIRMMT Schulich School

  2. Noisemes: Manual Annotation of Environmental Noise in Audio Streams Susanne Burger, Qin Jin, Peter F. Schulam, and Florian Metze

    E-print Network

    to technical complexity. High-quality, post- processed audio tracks from commercial videos where environmental#12;Noisemes: Manual Annotation of Environmental Noise in Audio Streams Susanne Burger, Qin Jin, Pittsburgh, PA 15213, USA {sburger,qjin,pschulam,fmetze}@cs.cmu.edu Abstract Audio information retrieval

  3. Autonomic QoS Optimization of Real-time Internet Audio using Loss Prediction and Stochastic Control

    E-print Network

    Wang, Yongge

    mechanisms are used to improve reliability and thus session quality. To achieve optimized receiver audio is to optimize receiver quality at real time using objective audio quality assessment. That may mean that certainAutonomic QoS Optimization of Real-time Internet Audio using Loss Prediction and Stochastic Control

  4. 76 FR 57923 - Establishment of Rules and Policies for the Satellite Digital Audio Radio Service in the 2310...

    Federal Register 2010, 2011, 2012, 2013, 2014

    2011-09-19

    ...sections. See Satellite Digital Audio Radio Service (SDARS) Second...Title: Satellite Digital Audio Radio Service (SDARS). Form...and Policies for the Digital Audio Radio Satellite Service in the...are necessary to ensure a high quality service to the public,...

  5. TEMPORAL SYNCHRONIZATION OF MULTIPLE AUDIO SIGNALS Julius Kammerl, Neil Birkbeck, Sasi Inguva, Damien Kelly, A. J. Crawford,

    E-print Network

    Cortes, Corinna

    degradation due to compression and noise artifacts im- pairs the audio quality leading to inconsistenciesTEMPORAL SYNCHRONIZATION OF MULTIPLE AUDIO SIGNALS Julius Kammerl, Neil Birkbeck, Sasi Inguva-based approaches for syn- chronizing multiple audio signals. The graphs are constructed atop the over

  6. IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, 2013. ACCEPTED FOR PUBLICATION 1 Active-Set Newton Algorithm for Overcomplete

    E-print Network

    Virtanen, Tuomas

    quality in less time. Index Terms--acoustic signal analysis, audio source separation, supervised sourceIEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, 2013. ACCEPTED FOR PUBLICATION 1 Active-Set Newton Algorithm for Overcomplete Non-Negative Representations of Audio Tuomas Virtanen

  7. Proc. of the 15th Int. Conference on Digital Audio Effects (DAFx-12), York, UK, September 17-21, 2012

    E-print Network

    Reiss, Josh

    subjective quality as those produced by the ex- pert engineer. 1. INTRODUCTION Multi-track audio mixingProc. of the 15th Int. Conference on Digital Audio Effects (DAFx-12), York, UK, September 17.maddams@gmail.com ABSTRACT Dynamic range compression is a nonlinear audio effect that re- duces the dynamic range of a signal

  8. ON THE IMPORTANCE OF "REAL" AUDIO DATA FOR MIR ALGORITHM EVALUATION AT THE NOTE-LEVEL A COMPARATIVE STUDY

    E-print Network

    Widmer, Gerhard

    data. This study investigates the effect of audio quality and source on the performance of two patches; a com- mercial high-quality sample library; and audio recordings made on a real (computerON THE IMPORTANCE OF "REAL" AUDIO DATA FOR MIR ALGORITHM EVALUATION AT THE NOTE

  9. Comparing the Effects of Classroom Audio-Recording and Video-Recording on Preservice Teachers' Reflection of Practice

    ERIC Educational Resources Information Center

    Bergman, Daniel

    2015-01-01

    This study examined the effects of audio and video self-recording on preservice teachers' written reflections. Participants (n = 201) came from a secondary teaching methods course and its school-based (clinical) fieldwork. The audio group (n[subscript A] = 106) used audio recorders to monitor their teaching in fieldwork placements; the video group…

  10. Machine Learning and Radiology

    PubMed Central

    Wang, Shijun; Summers, Ronald M.

    2012-01-01

    In this paper, we give a short introduction to machine learning and survey its applications in radiology. We focused on six categories of applications in radiology: medical image segmentation, registration, computer aided detection and diagnosis, brain function or activity analysis and neurological disease diagnosis from fMR images, content-based image retrieval systems for CT or MRI images, and text analysis of radiology reports using natural language processing (NLP) and natural language understanding (NLU). This survey shows that machine learning plays a key role in many radiology applications. Machine learning identifies complex patterns automatically and helps radiologists make intelligent decisions on radiology data such as conventional radiographs, CT, MRI, and PET images and radiology reports. In many applications, the performance of machine learning-based automatic detection and diagnosis systems has shown to be comparable to that of a well-trained and experienced radiologist. Technology development in machine learning and radiology will benefit from each other in the long run. Key contributions and common characteristics of machine learning techniques in radiology are discussed. We also discuss the problem of translating machine learning applications to the radiology clinical setting, including advantages and potential barriers. PMID:22465077

  11. The Basic Anaesthesia Machine

    PubMed Central

    Gurudatt, CL

    2013-01-01

    After WTG Morton's first public demonstration in 1846 of use of ether as an anaesthetic agent, for many years anaesthesiologists did not require a machine to deliver anaesthesia to the patients. After the introduction of oxygen and nitrous oxide in the form of compressed gases in cylinders, there was a necessity for mounting these cylinders on a metal frame. This stimulated many people to attempt to construct the anaesthesia machine. HEG Boyle in the year 1917 modified the Gwathmey's machine and this became popular as Boyle anaesthesia machine. Though a lot of changes have been made for the original Boyle machine still the basic structure remains the same. All the subsequent changes which have been brought are mainly to improve the safety of the patients. Knowing the details of the basic machine will make the trainee to understand the additional improvements. It is also important for every practicing anaesthesiologist to have a thorough knowledge of the basic anaesthesia machine for safe conduct of anaesthesia. PMID:24249876

  12. OPTICAM machine design

    NASA Astrophysics Data System (ADS)

    Liedes, Jyrki T.

    1992-01-01

    Rank Pneumo has worked with the Center of Optics Manufacturing to design a multiple-axis flexible machining center for spherical lens fabrication. The OPTICAM/SM prototype machine has been developed in cooperation with the Center's Manufacturing Advisory Board. The SM will generate, fine grind, pre-polish, and center a spherical lens surface in one setup sequence. Unique features of the design incorporate machine resident metrology to provide RQM (Real-time Quality Management) and closed-loop feedback control that corrects for lens thickness, diameter, and centering error. SPC (Statistical Process Control) software can compensate for process drift and QA data collection is provided without additional labor.

  13. Machine Tool Software

    NASA Technical Reports Server (NTRS)

    1988-01-01

    A NASA-developed software package has played a part in technical education of students who major in Mechanical Engineering Technology at William Rainey Harper College. Professor Hack has been using (APT) Automatically Programmed Tool Software since 1969 in his CAD/CAM Computer Aided Design and Manufacturing curriculum. Professor Hack teaches the use of APT programming languages for control of metal cutting machines. Machine tool instructions are geometry definitions written in APT Language to constitute a "part program." The part program is processed by the machine tool. CAD/CAM students go from writing a program to cutting steel in the course of a semester.

  14. The pendulum wave machine

    NASA Astrophysics Data System (ADS)

    Zetie, K. P.

    2015-05-01

    There are many examples on the internet of videos of ‘pendulum wave machines’ and how to make them (for example, www.instructables.com/id/Wave-Pendulum/). The machine is simply a set of pendula of different lengths which, when viewed end on, produce wave-like patterns from the positions of the bobs. These patterns change with time, with new patterns emerging as the bobs change phase. In this article, the physics of the machine is explored and explained, along with tips on how to build such a device.

  15. Effect of Audio Coaching on Correlation of Abdominal Displacement With Lung Tumor Motion

    SciTech Connect

    Nakamura, Mitsuhiro Narita, Yuichiro; Matsuo, Yukinori; Narabayashi, Masaru; Nakata, Manabu; Sawada, Akira; Mizowaki, Takashi; Nagata, Yasushi; Hiraoka, Masahiro

    2009-10-01

    Purpose: To assess the effect of audio coaching on the time-dependent behavior of the correlation between abdominal motion and lung tumor motion and the corresponding lung tumor position mismatches. Methods and Materials: Six patients who had a lung tumor with a motion range >8 mm were enrolled in the present study. Breathing-synchronized fluoroscopy was performed initially without audio coaching, followed by fluoroscopy with recorded audio coaching for multiple days. Two different measurements, anteroposterior abdominal displacement using the real-time positioning management system and superoinferior (SI) lung tumor motion by X-ray fluoroscopy, were performed simultaneously. Their sequential images were recorded using one display system. The lung tumor position was automatically detected with a template matching technique. The relationship between the abdominal and lung tumor motion was analyzed with and without audio coaching. Results: The mean SI tumor displacement was 10.4 mm without audio coaching and increased to 23.0 mm with audio coaching (p < .01). The correlation coefficients ranged from 0.89 to 0.97 with free breathing. Applying audio coaching, the correlation coefficients improved significantly (range, 0.93-0.99; p < .01), and the SI lung tumor position mismatches became larger in 75% of all sessions. Conclusion: Audio coaching served to increase the degree of correlation and make it more reproducible. In addition, the phase shifts between tumor motion and abdominal displacement were improved; however, all patients breathed more deeply, and the SI lung tumor position mismatches became slightly larger with audio coaching than without audio coaching.

  16. Digital audio watermarking using moment-preserving thresholding

    NASA Astrophysics Data System (ADS)

    Choi, DooSeop; Jung, Hae Kyung; Choi, Hyuk; Kim, Taejeong

    2007-09-01

    The Moment-Preserving Thresholding technique for digital images has been used in digital image processing for decades, especially in image binarization and image compression. Its main strength lies in that the binary values that the MPT produces as a result, called representative values, are usually unaffected when the signal being thresholded goes through a signal processing operation. The two representative values in MPT together with the threshold value are obtained by solving the system of the preservation equations for the first, second, and third moment. Relying on this robustness of the representative values to various signal processing attacks considered in the watermarking context, this paper proposes a new watermarking scheme for audio signals. The watermark is embedded in the root-sum-square (RSS) of the two representative values of each signal block using the quantization technique. As a result, the RSS values are modified by scaling the signal according to the watermark bit sequence under the constraint of inaudibility relative to the human psycho-acoustic model. We also address and suggest solutions to the problem of synchronization and power scaling attacks. Experimental results show that the proposed scheme maintains high audio quality and robustness to various attacks including MP3 compression, re-sampling, jittering, and, DA/AD conversion.

  17. Audio watermarking method robust against time- and frequency-fluctuation

    NASA Astrophysics Data System (ADS)

    Tachibana, Ryuki; Shimizu, Shuichi; Nakamura, Taiga; Kobayashi, Seiji

    2001-08-01

    In this paper, we describe an audio watermarking algorithm that can embed a multiple-bit message which is robust against wow-and-flutter, cropping, noise-addition, pitch-shift, and audio compressions such as MP3. The algorithm calculates and manipulates the magnitudes of segmented areas in the time-frequency plane of the content using short-term DFTs. The detection algorithm correlates the magnitudes with a pseudo-random array that maps to a two-dimensional area in the time-frequency plane. The two-dimensional array makes the watermark robust because, even when some portions of the content are heavily degraded, other portions of the content can match the pseudo-random array and contribute to watermark detection. Another key idea is manipulation of magnitudes. Because magnitudes are less influenced than phases by fluctuations of the analysis windows caused by random cropping, the watermark resists degradation. When signal transformation causes pitch fluctuations in the content, the frequencies of the pseudo-random array embedded in the content shift, and that causes a decrease in the volume of the watermark signal that still correctly overlaps with the corresponding pseudo-random array. To keep the overlapping area wide enough for successful watermark detection, the widths of the frequency subbands used for the detection segments should increase logarithmically as frequency increases. We theoretically and experimentally analyze the robustness of proposed algorithm against a variety of signal degradations.

  18. Separate Mechanisms for Audio-Tactile Pitch and Loudness Interactions

    PubMed Central

    Yau, Jeffrey M.; Weber, Alison I.; Bensmaia, Sliman J.

    2010-01-01

    A major goal in perceptual neuroscience is to understand how signals from different sensory modalities are combined to produce stable and coherent representations. We previously investigated interactions between audition and touch, motivated by the fact that both modalities are sensitive to environmental oscillations. In our earlier study, we characterized the effect of auditory distractors on tactile frequency and intensity perception. Here, we describe the converse experiments examining the effect of tactile distractors on auditory processing. Because the two studies employ the same psychophysical paradigm, we combined their results for a comprehensive view of how auditory and tactile signals interact and how these interactions depend on the perceptual task. Together, our results show that temporal frequency representations are perceptually linked regardless of the attended modality. In contrast, audio-tactile loudness interactions depend on the attended modality: Tactile distractors influence judgments of auditory intensity, but judgments of tactile intensity are impervious to auditory distraction. Lastly, we show that audio-tactile loudness interactions depend critically on stimulus timing, while pitch interactions do not. These results reveal that auditory and tactile inputs are combined differently depending on the perceptual task. That distinct rules govern the integration of auditory and tactile signals in pitch and loudness perception implies that the two are mediated by separate neural mechanisms. These findings underscore the complexity and specificity of multisensory interactions. PMID:21887147

  19. A compact electroencephalogram recording device with integrated audio stimulation system

    NASA Astrophysics Data System (ADS)

    Paukkunen, Antti K. O.; Kurttio, Anttu A.; Leminen, Miika M.; Sepponen, Raimo E.

    2010-06-01

    A compact (96×128×32 mm3, 374 g), battery-powered, eight-channel electroencephalogram recording device with an integrated audio stimulation system and a wireless interface is presented. The recording device is capable of producing high-quality data, while the operating time is also reasonable for evoked potential studies. The effective measurement resolution is about 4 nV at 200 Hz sample rate, typical noise level is below 0.7 ?Vrms at 0.16-70 Hz, and the estimated operating time is 1.5 h. An embedded audio decoder circuit reads and plays wave sound files stored on a memory card. The activities are controlled by an 8 bit main control unit which allows accurate timing of the stimuli. The interstimulus interval jitter measured is less than 1 ms. Wireless communication is made through bluetooth and the data recorded are transmitted to an external personal computer (PC) interface in real time. The PC interface is implemented with LABVIEW® and in addition to data acquisition it also allows online signal processing, data storage, and control of measurement activities such as contact impedance measurement, for example. The practical application of the device is demonstrated in mismatch negativity experiment with three test subjects.

  20. Audio watermarking technologies for automatic cue sheet generation systems

    NASA Astrophysics Data System (ADS)

    Caccia, Giuseppe; Lancini, Rosa C.; Pascarella, Annalisa; Tubaro, Stefano; Vicario, Elena

    2001-08-01

    Usually watermark is used as a way for hiding information on digital media. The watermarked information may be used to allow copyright protection or user and media identification. In this paper we propose a watermarking scheme for digital audio signals that allow automatic identification of musical pieces transmitted in TV broadcasting programs. In our application the watermark must be, obviously, imperceptible to the users, should be robust to standard TV and radio editing and have a very low complexity. This last item is essential to allow a software real-time implementation of the insertion and detection of watermarks using only a minimum amount of the computation power of a modern PC. In the proposed method the input audio sequence is subdivided in frames. For each frame a watermark spread spectrum sequence is added to the original data. A two steps filtering procedure is used to generate the watermark from a Pseudo-Noise (PN) sequence. The filters approximate respectively the threshold and the frequency masking of the Human Auditory System (HAS). In the paper we discuss first the watermark embedding system then the detection approach. The results of a large set of subjective tests are also presented to demonstrate the quality and robustness of the proposed approach.

  1. Audio-tactile integration and the influence of musical training.

    PubMed

    Kuchenbuch, Anja; Paraskevopoulos, Evangelos; Herholz, Sibylle C; Pantev, Christo

    2014-01-01

    Perception of our environment is a multisensory experience; information from different sensory systems like the auditory, visual and tactile is constantly integrated. Complex tasks that require high temporal and spatial precision of multisensory integration put strong demands on the underlying networks but it is largely unknown how task experience shapes multisensory processing. Long-term musical training is an excellent model for brain plasticity because it shapes the human brain at functional and structural levels, affecting a network of brain areas. In the present study we used magnetoencephalography (MEG) to investigate how audio-tactile perception is integrated in the human brain and if musicians show enhancement of the corresponding activation compared to non-musicians. Using a paradigm that allowed the investigation of combined and separate auditory and tactile processing, we found a multisensory incongruency response, generated in frontal, cingulate and cerebellar regions, an auditory mismatch response generated mainly in the auditory cortex and a tactile mismatch response generated in frontal and cerebellar regions. The influence of musical training was seen in the audio-tactile as well as in the auditory condition, indicating enhanced higher-order processing in musicians, while the sources of the tactile MMN were not influenced by long-term musical training. Consistent with the predictive coding model, more basic, bottom-up sensory processing was relatively stable and less affected by expertise, whereas areas for top-down models of multisensory expectancies were modulated by training. PMID:24465675

  2. Audio-Tactile Integration and the Influence of Musical Training

    PubMed Central

    Kuchenbuch, Anja; Paraskevopoulos, Evangelos; Herholz, Sibylle C.; Pantev, Christo

    2014-01-01

    Perception of our environment is a multisensory experience; information from different sensory systems like the auditory, visual and tactile is constantly integrated. Complex tasks that require high temporal and spatial precision of multisensory integration put strong demands on the underlying networks but it is largely unknown how task experience shapes multisensory processing. Long-term musical training is an excellent model for brain plasticity because it shapes the human brain at functional and structural levels, affecting a network of brain areas. In the present study we used magnetoencephalography (MEG) to investigate how audio-tactile perception is integrated in the human brain and if musicians show enhancement of the corresponding activation compared to non-musicians. Using a paradigm that allowed the investigation of combined and separate auditory and tactile processing, we found a multisensory incongruency response, generated in frontal, cingulate and cerebellar regions, an auditory mismatch response generated mainly in the auditory cortex and a tactile mismatch response generated in frontal and cerebellar regions. The influence of musical training was seen in the audio-tactile as well as in the auditory condition, indicating enhanced higher-order processing in musicians, while the sources of the tactile MMN were not influenced by long-term musical training. Consistent with the predictive coding model, more basic, bottom-up sensory processing was relatively stable and less affected by expertise, whereas areas for top-down models of multisensory expectancies were modulated by training. PMID:24465675

  3. Information-Driven Active Audio-Visual Source Localization

    PubMed Central

    Schult, Niclas; Reineking, Thomas; Kluss, Thorsten; Zetzsche, Christoph

    2015-01-01

    We present a system for sensorimotor audio-visual source localization on a mobile robot. We utilize a particle filter for the combination of audio-visual information and for the temporal integration of consecutive measurements. Although the system only measures the current direction of the source, the position of the source can be estimated because the robot is able to move and can therefore obtain measurements from different directions. These actions by the robot successively reduce uncertainty about the source’s position. An information gain mechanism is used for selecting the most informative actions in order to minimize the number of actions required to achieve accurate and precise position estimates in azimuth and distance. We show that this mechanism is an efficient solution to the action selection problem for source localization, and that it is able to produce precise position estimates despite simplified unisensory preprocessing. Because of the robot’s mobility, this approach is suitable for use in complex and cluttered environments. We present qualitative and quantitative results of the system’s performance and discuss possible areas of application. PMID:26327619

  4. FORUM ON MACHINE TRANSLATION What Should Machine Translation Be?

    E-print Network

    FORUM ON MACHINE TRANSLATION What Should Machine Translation Be? John S. White Siemens Information After a considerable hiatus of interest and funding, machine translation has come in recent years of the most visible representations of natural language processing to the outside world. Machine translation

  5. INTRODUCTION TO MACHINING 1.2 MILLING MACHINE

    E-print Network

    Sun, Yu

    INTRODUCTION TO MACHINING 1.2 MILLING MACHINE: Milling is a fabrication process during which a work, and the rotating drill bit is moved, vertically, into and out of the work-piece. The process of milling is capable(s) being used and the sophistication and type of the milling machine. Milling machines are referred to as 3

  6. FORUM ON MACHINE TRANSLATION Machine Translation already does Work

    E-print Network

    research tools seems quite uncontentious. Now, just as machine translation is not a single in- divisibleFORUM ON MACHINE TRANSLATION Machine Translation already does Work Margaret King ISSCO 54, rte des "Does machine translation work is that the question itself is ill- posed. It takes for granted

  7. Universal Developer Platform for MachinePlatform for Machine

    E-print Network

    1 Universal Developer Platform for MachinePlatform for Machine Control Applications Hansjoerg will need some things: · Senses · Means June 24-26, 2008 ETH Zurich #12;2 And for Machine Control? June 24 ETH Zurich Conclusion: big effort for things that have little to do with Machine Control Multi

  8. 16. Interior, Machine Shop, Roundhouse Machine Shop Extension, Southern Pacific ...

    Library of Congress Historic Buildings Survey, Historic Engineering Record, Historic Landscapes Survey

    16. Interior, Machine Shop, Roundhouse Machine Shop Extension, Southern Pacific Railroad Carlin Shops, view to south (90mm lens). Note the large segmental-arched doorway to move locomotives in and out of Machine Shop. - Southern Pacific Railroad, Carlin Shops, Roundhouse Machine Shop Extension, Foot of Sixth Street, Carlin, Elko County, NV

  9. Machine learning blocks

    E-print Network

    Collazo Santiago, Bryan Omar

    2015-01-01

    This work presents MLBlocks, a machine learning system that lets data scientists explore the space of modeling techniques in a very easy and efficient manner. We show how the system is very general in the sense that virtually ...

  10. Machine For Grinding Fog

    E-print Network

    Ward, Nicholas Adam

    2010-04-30

    Machine For Grinding Fog, an exhibition of stop motion animated films, revolves around three primary concerns: relationships, psychological or subjective reality, and relativity. Rooted within concerns of faith, personal ethics, and a re...

  11. Tunnel boring machine

    SciTech Connect

    Snyder, L. L.

    1985-07-09

    A tunnel boring machine for controlled boring of a curvilinear tunnel including a rotating cutter wheel mounted on the forward end of a thrust cylinder assembly having a central longitudinal axis aligned with the cutter wheel axis of rotation; the thrust cylinder assembly comprising a cylinder barrel and an extendable and retractable thrust arm received therein. An anchoring assembly is pivotally attached to the rear end of the cylinder barrel for anchoring the machine during a cutting stroke and providing a rear end pivot axis during curved cutting strokes. A pair of laterally extending, extendable and retractable arms are fixedly mounted at a forward portion of the cylinder barrel for providing lateral displacement in a laterally curved cutting mode and for anchoring the machine between cutting strokes and during straight line boring. Forward and rear transverse displacement and support assemblies are provided to facilitate cutting in a transversely curved cutting mode and to facilitate machine movement between cutting strokes.

  12. Design of Vitrification Machine

    E-print Network

    McDonald, Todd William

    2009-11-30

    and unreliable process. Moreover, there is a lack of standardization in the methods for preparing cells and biological material for the vitrification process. The purpose of the Vitrification Machine is to greatly simplify the process by making it faster, more...

  13. Machine Translation for Twitter 

    E-print Network

    Jehl, Laura Elisabeth

    2010-11-24

    We carried out a study in which we explored the feasibility of machine translation for Twitter for the language pair English and German. As a first step we created a small bilingual corpus of 1,000 tweets. Using this ...

  14. Laser machining of ceramic

    SciTech Connect

    Laudel, A.

    1980-01-01

    The Kansas City Division of The Bendix Corporation manufactures hybrid microcircuits (HMCs) using both thin film and thick film technologies. Laser machining is used to contour the ceramic substrates and to drill holes in the ceramic for frontside-backside interconnections (vias) and holes for mounting components. A 1000 W CO/sub 2/ type laser is used. The laser machining process, and methods used for removing protruding debris and debris from holes, for cleaning the machined surfaces, and for refiring are described. The laser machining process described consistently produces vias, component holes and contours with acceptable surface quality, hole locations, diameter, flatness and metallization adhesion. There are no cracks indicated by dipping in fluorescent dye penetrant and the substances are resistant to repeated thermal shock.

  15. Zigzags in Turing Machines

    NASA Astrophysics Data System (ADS)

    Gajardo, Anahí; Guillon, Pierre

    We study one-head machines through symbolic and topological dynamics. In particular, a subshift is associated to the subshift, and we are interested in its complexity in terms of realtime recognition. We emphasize the class of one-head machines whose subshift can be recognized by a deterministic pushdown automaton. We prove that this class corresponds to particular restrictions on the head movement, and to equicontinuity in associated dynamical systems.

  16. Metalworking and machining fluids

    DOEpatents

    Erdemir, Ali (Naperville, IL); Sykora, Frank (Caledon, ON, CA); Dorbeck, Mark (Brighton, MI)

    2010-10-12

    Improved boron-based metal working and machining fluids. Boric acid and boron-based additives that, when mixed with certain carrier fluids, such as water, cellulose and/or cellulose derivatives, polyhydric alcohol, polyalkylene glycol, polyvinyl alcohol, starch, dextrin, in solid and/or solvated forms result in improved metalworking and machining of metallic work pieces. Fluids manufactured with boric acid or boron-based additives effectively reduce friction, prevent galling and severe wear problems on cutting and forming tools.

  17. Doubly fed induction machine

    DOEpatents

    Skeist, S. Merrill; Baker, Richard H.

    2005-10-11

    An electro-mechanical energy conversion system coupled between an energy source and an energy load including an energy converter device having a doubly fed induction machine coupled between the energy source and the energy load to convert the energy from the energy source and to transfer the converted energy to the energy load and an energy transfer multiplexer coupled to the energy converter device to control the flow of power or energy through the doubly fed induction machine.

  18. Human-machine interactions

    DOEpatents

    Forsythe, J. Chris (Sandia Park, NM); Xavier, Patrick G. (Albuquerque, NM); Abbott, Robert G. (Albuquerque, NM); Brannon, Nathan G. (Albuquerque, NM); Bernard, Michael L. (Tijeras, NM); Speed, Ann E. (Albuquerque, NM)

    2009-04-28

    Digital technology utilizing a cognitive model based on human naturalistic decision-making processes, including pattern recognition and episodic memory, can reduce the dependency of human-machine interactions on the abilities of a human user and can enable a machine to more closely emulate human-like responses. Such a cognitive model can enable digital technology to use cognitive capacities fundamental to human-like communication and cooperation to interact with humans.

  19. Sealing intersecting vane machines

    DOEpatents

    Martin, Jedd N. (Providence, RI); Chomyszak, Stephen M. (Attleboro, MA)

    2007-06-05

    The invention provides a toroidal intersecting vane machine incorporating intersecting rotors to form primary and secondary chambers whose porting configurations minimize friction and maximize efficiency. Specifically, it is an object of the invention to provide a toroidal intersecting vane machine that greatly reduces the frictional losses through intersecting surfaces without the need for external gearing by modifying the width of one or both tracks at the point of intermeshing. The inventions described herein relate to these improvements.

  20. Sealing intersecting vane machines

    DOEpatents

    Martin, Jedd N.; Chomyszak, Stephen M.

    2005-06-07

    The invention provides a toroidal intersecting vane machine incorporating intersecting rotors to form primary and secondary chambers whose porting configurations minimize friction and maximize efficiency. Specifically, it is an object of the invention to provide a toroidal intersecting vane machine that greatly reduces the frictional losses through intersecting surfaces without the need for external gearing by modifying the width of one or both tracks at the point of intermeshing. The inventions described herein relate to these improvements.

  1. Monitoring frog communities: An application of machine learning

    SciTech Connect

    Taylor, A.; Watson, G.; Grigg, G.; McCallum, H.

    1996-12-31

    Automatic recognition of animal vocalizations would be a valuable tool for a variety of biological research and environmental monitoring applications. We report the development of a software system which can recognize the vocalizations of 22 species of frogs which occur in an area of northern Australia. This software system will be used in unattended operation to monitor the effect on frog populations of the introduced Cane Toad. The system is based around classification of local peaks in the spectrogram of the audio signal using Quinlan`s machine learning system, C4.5. Unreliable identifications of peaks are aggregated together using a hierarchical structure of segments based on the typical temporal vocalization species` patterns. This produces robust system performance.

  2. Could a machine think

    SciTech Connect

    Churchland, P.M.; Churchland, P.S. )

    1990-01-01

    There are many reasons for saying yes. One of the earliest and deepest reason lay in two important results in computational theory. The first was Church's thesis, which states that every effectively computable function is recursively computable. The second important result was Alan M. Turing's demonstration that any recursively computable function can be computed in finite time by a maximally simple sort of symbol-manipulating machine that has come to be called a universal Turing machine. This machine is guided by a set of recursively applicable rules that are sensitive to the identity, order and arrangement of the elementary symbols it encounters as input. The authors reject the Turing test as a sufficient condition for conscious intelligence. They base their position of the specific behavioral failures of the classical SM machines and on the specific virtues of machines with a more brain-like architecture. These contrasts show that certain computational strategies have vast and decisive advantages over others where typical cognitive tasks are concerned, advantages that are empirically inescapable. Clearly, the brain is making systematic use of these computational advantage. But it need not be the only physical system capable of doing so. Artificial intelligence, in a nonbiological but massively parallel machine, remain a compelling and discernible prospect.

  3. Machine Learning Srihari Decision Trees

    E-print Network

    Machine Learning Srihari 1 Decision Trees Sargur Srihari srihari@cedar.buffalo.edu #12;Machine the unobserved examples · Inductive bias is a preference for small trees over large trees #12;Machine Learning #12;Machine Learning Srihari 4 Learning Data Day Outlook Temp Humidity Wind PlayTennis D1 Sunny Hot

  4. Machine Learning Thomas G. Dietterich

    E-print Network

    Machine Learning Thomas G. Dietterich Department of Computer Science Oregon State University Corvallis, OR 97331 1 Introduction Machine Learning is the study of methods for programming computers a computer system. A machine learning system can study recorded data and subsequent machine failures

  5. Harvesting Machine Census 1999 & 2001

    E-print Network

    1 Harvesting Machine Census 1999 & 2001 231 Corstorphine Road Edinburgh EH12 7AT www.forestry.gov.uk FCTN001 SUMMARY This Technical Note contains information on the 1999 and 2001 harvesting machine machines, converted forwarders, etc., account for the remaining machines. In the 2001 census, 65

  6. WISKUNNEND WISKE DE MEESTERLIJKE MACHINE

    E-print Network

    Goelzer, Heiko

    WISKUNNEND WISKE DE MEESTERLIJKE MACHINE ©2013,StandaardUitgeverij,Antwerpen,België Lara werkt studio een machine die pagina's kan sorteren op één van de 5 gecodeerde cijfers: wanneer je de machine bv ingestelde plaats hetzelfde cijfer hebben, dan behoudt de machine hun onderlinge volgorde. a) Help Lara met

  7. Machine Learning Srihari Neural Networks

    E-print Network

    Machine Learning Srihari Neural Networks Introduction Sargur Srihari #12;Machine Learning in Neural Networks 6. Mixture Density Networks 7. Bayesian Neural Networks 2 #12;Machine Learning xi + w j 0 (1) + wk0 (2) #12;Machine Learning Srihari SVM versus Neural Networks

  8. Non-traditional machining techniques

    SciTech Connect

    Day, Robert D; Fierro, Frank; Garcia, Felix P; Hatch, Douglass J; Randolph, Randall B; Reardon, Patrick T; Rivera, Gerald

    2008-01-01

    During the course of machining targets for various experiments it sometimes becomes necessary to adapt fixtures or machines, which are designed for one function, to another function. When adapting a machine or fixture is not adequate, it may be necessary to acquire a machine specifically designed to produce the component required. In addition to the above scenarios, the features of a component may dictate that multi-step machining processes are necessary to produce the component. This paper discusses the machining of four components where adaptation, specialized machine design, or multi-step processes were necessary to produce the components.

  9. To be published in Proceedings of the ACM Conference on Computer-Supported Cooperative Work (CSCW'96). Thunderwire: A Field Study of an Audio-Only Media Space

    E-print Network

    Ackerman, Mark S.

    using an audio-only media space. This media space, called Thunderwire, combined high-quality audio space over several months. This media space system, called Thunderwire, combined high-quality audio'96). Thunderwire: A Field Study of an Audio-Only Media Space Debby Hindus1, Mark S. Ackerman2, Scott

  10. IEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 16, NO. 8, NOVEMBER 2008 1373 A New Model-Based Algorithm for Optimizing

    E-print Network

    Richard, Gaël

    THE MPEG-4 Advanced Audio Coder (AAC) is the latest international standard for high-quality lossy audioIEEE TRANSACTIONS ON AUDIO, SPEECH, AND LANGUAGE PROCESSING, VOL. 16, NO. 8, NOVEMBER 2008 1373-Advanced Audio Coder (AAC) in MS-stereo mode is presented. This algorithm is an extension to stereo sig- nals

  11. Design and implementation of a two-way real-time communication system for audio over CATV networks

    NASA Astrophysics Data System (ADS)

    Cho, Choong Sang; Oh, Yoo Rhee; Lee, Young Han; Kim, Hong Kook

    2007-09-01

    In this paper, we design and implement a two-way real-time communication system for audio over cable television (CATV) networks to provide an audio-based interaction between the CATV broadcasting station and CATV subscribers. The two-way real-time communication system consists of a real-time audio encoding/decoding module, a payload formatter based on a transmission control protocol/Internet protocol (TCP/IP), and a cable network. At the broadcasting station, audio signals from a microphone are encoded by an audio codec that is implemented using a digital signal processor (DSP), where the MPEG-2 Layer II audio codec is used for the audio codec and TMS320C6416 is used for a DSP. Next, a payload formatter constructs a TCP/IP packet from an audio bitstream for transmission to a cable modem. Another payload formatter at the subscriber unpacks the TCP/IP packet decoded from the cable modem into audio bitstream. This bitstream is decoded by the MPEG-2 Layer II audio decoder. Finally the decoded audio signals are played out to the speaker. We confirmed that the system worked in real-time, with a measured delay of around 150 ms including the algorithmic and processing time delays.

  12. TECHNICAL NOTE: Portable audio electronics for impedance-based measurements in microfluidics

    NASA Astrophysics Data System (ADS)

    Wood, Paul; Sinton, David

    2010-08-01

    We demonstrate the use of audio electronics-based signals to perform on-chip electrochemical measurements. Cell phones and portable music players are examples of consumer electronics that are easily operated and are ubiquitous worldwide. Audio output (play) and input (record) signals are voltage based and contain frequency and amplitude information. A cell phone, laptop soundcard and two compact audio players are compared with respect to frequency response; the laptop soundcard provides the most uniform frequency response, while the cell phone performance is found to be insufficient. The audio signals in the common portable music players and laptop soundcard operate in the range of 20 Hz to 20 kHz and are found to be applicable, as voltage input and output signals, to impedance-based electrochemical measurements in microfluidic systems. Validated impedance-based measurements of concentration (0.1-50 mM), flow rate (2-120 µL min-1) and particle detection (32 µm diameter) are demonstrated. The prevailing, lossless, wave audio file format is found to be suitable for data transmission to and from external sources, such as a centralized lab, and the cost of all hardware (in addition to audio devices) is ~10 USD. The utility demonstrated here, in combination with the ubiquitous nature of portable audio electronics, presents new opportunities for impedance-based measurements in portable microfluidic systems.

  13. Laboratory and in-flight experiments to evaluate 3-D audio display technology

    NASA Technical Reports Server (NTRS)

    Ericson, Mark; Mckinley, Richard; Kibbe, Marion; Francis, Daniel

    1994-01-01

    Laboratory and in-flight experiments were conducted to evaluate 3-D audio display technology for cockpit applications. A 3-D audio display generator was developed which digitally encodes naturally occurring direction information onto any audio signal and presents the binaural sound over headphones. The acoustic image is stabilized for head movement by use of an electromagnetic head-tracking device. In the laboratory, a 3-D audio display generator was used to spatially separate competing speech messages to improve the intelligibility of each message. Up to a 25 percent improvement in intelligibility was measured for spatially separated speech at high ambient noise levels (115 dB SPL). During the in-flight experiments, pilots reported that spatial separation of speech communications provided a noticeable improvement in intelligibility. The use of 3-D audio for target acquisition was also investigated. In the laboratory, 3-D audio enabled the acquisition of visual targets in about two seconds average response time at 17 degrees accuracy. During the in-flight experiments, pilots correctly identified ground targets 50, 75, and 100 percent of the time at separation angles of 12, 20, and 35 degrees, respectively. In general, pilot performance in the field with the 3-D audio display generator was as expected, based on data from laboratory experiments.

  14. Audio-vestibular evaluation in patients with Behçet's syndrome.

    PubMed

    Evereklioglu, C; Cokkeser, Y; Doganay, S; Er, H; Kizilay, A

    2001-09-01

    A prospective controlled clinical study was carried out at the Department of Ophthalmology and ENT, Inönü University Medical Faculty, Turgut Ozal Medical Center, Research Hospital, to evaluate the audio-vestibular involvement in patients with Behçet's syndrome compared with controls. Twenty-five consecutive patients with Behçet's syndrome (mean age +/- SD, 34.96 +/- 8.50) and 20 age- and sex-matched healthy volunteers (hospital staff) as control subjects (mean age +/- SD, 34.45 +/- 9.16) were included in this study. Behçet's patients were divided into two groups according to the number of criteria, complete (all four major criteria) and incomplete (three major criteria without ocular involvement). The groups were compared with each other or controls regarding inner ear involvement. Audiometric pure-tone thresholds at 125 to 8000 Hz were obtained in all subjects in both groups, and pure tone average (PTA) hearing thresholds were calculated for the middle, high and low frequencies. In addition, short increment sensitivity index (SISI), tone decay and BERA examinations were performed in all Behçet's patients. Sensorineural hearing loss (SNHL) was present in six of 25 patients with Behçet's syndrome. Two Behçet's patients had unilateral total SNHL, two had bilateral moderate level SNHL, one had bilateral low-frequency SNHL and one bilateral high frequency SNHL. In two, BERA, and in five SISI, examination disclosed inner ear involvement. In control subjects, the past medical history was normal and there was no consistent audio-vestibular complaint. Their PTA thresholds were all in the normal range. Otoscopic examination findings were normal, with intact, mobile tympanic membranes in both groups. The present study showed that audio-vestibular involvement is not infrequent in Behçet's syndrome compared with age- and sex-matched healthy controls, and it is under-estimated. All Behçet's patients should regularly be followed by an otolaryngologists and be given information about the possibility of inner ear involvement. According to our results, hearing loss occurs more often in older patients and also in the complete form of Behçet's syndrome. PMID:11564295

  15. Small Weakly Universal Turing Machines

    NASA Astrophysics Data System (ADS)

    Neary, Turlough; Woods, Damien

    We give small universal Turing machines with state-symbol pairs of (6,2), (3,3) and (2,4). These machines are weakly universal, which means that they have an infinitely repeated word to the left of their input and another to the right. They simulate Rule 110 and are currently the smallest known weakly universal Turing machines. Despite their small size these machines are efficient polynomial time simulators of Turing machines.

  16. Method for reading sensors and controlling actuators using audio interfaces of mobile devices.

    PubMed

    Aroca, Rafael V; Burlamaqui, Aquiles F; Gonçalves, Luiz M G

    2012-01-01

    This article presents a novel closed loop control architecture based on audio channels of several types of computing devices, such as mobile phones and tablet computers, but not restricted to them. The communication is based on an audio interface that relies on the exchange of audio tones, allowing sensors to be read and actuators to be controlled. As an application example, the presented technique is used to build a low cost mobile robot, but the system can also be used in a variety of mechatronics applications and sensor networks, where smartphones are the basic building blocks. PMID:22438726

  17. An overview of statistical-model-based techniques for audio restoration

    NASA Astrophysics Data System (ADS)

    Wolfe, Patrick J.

    2001-05-01

    This presentation will describe the state of the art in model-based approaches to audio restoration, with an emphasis on results and an explanation of theory suitable for the nonspecialist. An overview of work conducted over the past several years by members of the Cambridge Signal Processing Group will be presented, including techniques for the detection, correction, and interpolation of degraded or missing audio data. The latest advances in a fully Bayesian statistical methodology for modeling of music and speech signals will also be detailed, including a common unifying framework for the treatment of both global and localized audio degradations.

  18. Method for Reading Sensors and Controlling Actuators Using Audio Interfaces of Mobile Devices

    PubMed Central

    Aroca, Rafael V.; Burlamaqui, Aquiles F.; Gonçalves, Luiz M. G.

    2012-01-01

    This article presents a novel closed loop control architecture based on audio channels of several types of computing devices, such as mobile phones and tablet computers, but not restricted to them. The communication is based on an audio interface that relies on the exchange of audio tones, allowing sensors to be read and actuators to be controlled. As an application example, the presented technique is used to build a low cost mobile robot, but the system can also be used in a variety of mechatronics applications and sensor networks, where smartphones are the basic building blocks. PMID:22438726

  19. The Brussels Mood Inductive Audio Stories (MIAS) database.

    PubMed

    Bertels, Julie; Deliens, Gaétane; Peigneux, Philippe; Destrebecqz, Arnaud

    2014-12-01

    Through this study, we aimed to validate a new tool for inducing moods in experimental contexts. Five audio stories with sad, joyful, frightening, erotic, or neutral content were presented to 60 participants (33 women, 27 men) in a within-subjects design, each for about 10 min. Participants were asked (1) to report their moods before and after listening to each story, (2) to assess the emotional content of the excerpts on various emotional scales, and (3) to rate their level of projection into the stories. The results confirmed our a priori emotional classification. The emotional stories were effective in inducing the desired mood, with no difference found between male and female participants. These stories therefore constitute a valuable corpus for inducing moods in French-speaking participants, and they are made freely available for use in scientific research. PMID:24519495

  20. Audio-visual speech perception: a developmental ERP investigation

    PubMed Central

    Knowland, Victoria CP; Mercure, Evelyne; Karmiloff-Smith, Annette; Dick, Fred; Thomas, Michael SC

    2014-01-01

    Being able to see a talking face confers a considerable advantage for speech perception in adulthood. However, behavioural data currently suggest that children fail to make full use of these available visual speech cues until age 8 or 9. This is particularly surprising given the potential utility of multiple informational cues during language learning. We therefore explored this at the neural level. The event-related potential (ERP) technique has been used to assess the mechanisms of audio-visual speech perception in adults, with visual cues reliably modulating auditory ERP responses to speech. Previous work has shown congruence-dependent shortening of auditory N1/P2 latency and congruence-independent attenuation of amplitude in the presence of auditory and visual speech signals, compared to auditory alone. The aim of this study was to chart the development of these well-established modulatory effects over mid-to-late childhood. Experiment 1 employed an adult sample to validate a child-friendly stimulus set and paradigm by replicating previously observed effects of N1/P2 amplitude and latency modulation by visual speech cues; it also revealed greater attenuation of component amplitude given incongruent audio-visual stimuli, pointing to a new interpretation of the amplitude modulation effect. Experiment 2 used the same paradigm to map cross-sectional developmental change in these ERP responses between 6 and 11 years of age. The effect of amplitude modulation by visual cues emerged over development, while the effect of latency modulation was stable over the child sample. These data suggest that auditory ERP modulation by visual speech represents separable underlying cognitive processes, some of which show earlier maturation than others over the course of development. PMID:24176002