Light Weight MP3 Watermarking Method for Mobile Terminals
NASA Astrophysics Data System (ADS)
Takagi, Koichi; Sakazawa, Shigeyuki; Takishima, Yasuhiro
This paper proposes a novel MP3 watermarking method which is applicable to a mobile terminal with limited computational resources. Considering that in most cases the embedded information is copyright information or metadata, which should be extracted before playing back audio contents, the watermark detection process should be executed at high speed. However, when conventional methods are used with a mobile terminal, it takes a considerable amount of time to detect a digital watermark. This paper focuses on scalefactor manipulation to enable high speed watermark embedding/detection for MP3 audio and also proposes the manipulation method which minimizes audio quality degradation adaptively. Evaluation tests showed that the proposed method is capable of embedding 3 bits/frame information without degrading audio quality and detecting it at very high speed. Finally, this paper describes application examples for authentication with a digital signature.
ERIC Educational Resources Information Center
Desmarais, Norman
1991-01-01
Reviews current developments in multimedia computing for both the business and consumer markets, including interactive multimedia players; compact disc-interactive (CD-I), including levels of audio quality, various video specifications and visual effects, and software; digital video interactive (DVI); and multimedia personal computers. (LRW)
Implementing Audio-CASI on Windows’ Platforms
Cooley, Philip C.; Turner, Charles F.
2011-01-01
Audio computer-assisted self interviewing (Audio-CASI) technologies have recently been shown to provide important and sometimes dramatic improvements in the quality of survey measurements. This is particularly true for measurements requiring respondents to divulge highly sensitive information such as their sexual, drug use, or other sensitive behaviors. However, DOS-based Audio-CASI systems that were designed and adopted in the early 1990s have important limitations. Most salient is the poor control they provide for manipulating the video presentation of survey questions. This article reports our experiences adapting Audio-CASI to Microsoft Windows 3.1 and Windows 95 platforms. Overall, our Windows-based system provided the desired control over video presentation and afforded other advantages including compatibility with a much wider array of audio devices than our DOS-based Audio-CASI technologies. These advantages came at the cost of increased system requirements --including the need for both more RAM and larger hard disks. While these costs will be an issue for organizations converting large inventories of PCS to Windows Audio-CASI today, this will not be a serious constraint for organizations and individuals with small inventories of machines to upgrade or those purchasing new machines today. PMID:22081743
A La Carts: You Want Wireless Mobility? Have a COW
ERIC Educational Resources Information Center
Villano, Matt
2006-01-01
Computers on wheels, or COWs, combine the wireless technology of today with the audio/visual carts of yesteryear for an entirely new spin on mobility. Increasingly used by districts with laptop computing initiatives, COWs are among the hottest high-tech sellers in schools today, according to market research firm Quality Education Data. In this…
New Integrated Video and Graphics Technology: Digital Video Interactive.
ERIC Educational Resources Information Center
Optical Information Systems, 1987
1987-01-01
Describes digital video interactive (DVI), a new technology which combines the interactivity of the graphics capabilities in personal computers with the realism of high-quality motion video and multitrack audio in an all-digital integrated system. (MES)
NASA Astrophysics Data System (ADS)
Esquef, Paulo A. A.
The first reproducible recording of human voice was made in 1877 on a tinfoil cylinder phonograph devised by Thomas A. Edison. Since then, much effort has been expended to find better ways to record and reproduce sounds. By the mid-1920s, the first electrical recordings appeared and gradually took over purely acoustic recordings. The development of electronic computers, in conjunction with the ability to record data onto magnetic or optical media, culminated in the standardization of compact disc format in 1980. Nowadays, digital technology is applied to several audio applications, not only to improve the quality of modern and old recording/reproduction techniques, but also to trade off sound quality for less storage space and less taxing transmission capacity requirements.
Defraene, Bruno; van Waterschoot, Toon; Diehl, Moritz; Moonen, Marc
2016-07-01
Subjective audio quality evaluation experiments have been conducted to assess the performance of embedded-optimization-based precompensation algorithms for mitigating perceptible linear and nonlinear distortion in audio signals. It is concluded with statistical significance that the perceived audio quality is improved by applying an embedded-optimization-based precompensation algorithm, both in case (i) nonlinear distortion and (ii) a combination of linear and nonlinear distortion is present. Moreover, a significant positive correlation is reported between the collected subjective and objective PEAQ audio quality scores, supporting the validity of using PEAQ to predict the impact of linear and nonlinear distortion on the perceived audio quality.
Predicting the Overall Spatial Quality of Automotive Audio Systems
NASA Astrophysics Data System (ADS)
Koya, Daisuke
The spatial quality of automotive audio systems is often compromised due to their unideal listening environments. Automotive audio systems need to be developed quickly due to industry demands. A suitable perceptual model could evaluate the spatial quality of automotive audio systems with similar reliability to formal listening tests but take less time. Such a model is developed in this research project by adapting an existing model of spatial quality for automotive audio use. The requirements for the adaptation were investigated in a literature review. A perceptual model called QESTRAL was reviewed, which predicts the overall spatial quality of domestic multichannel audio systems. It was determined that automotive audio systems are likely to be impaired in terms of the spatial attributes that were not considered in developing the QESTRAL model, but metrics are available that might predict these attributes. To establish whether the QESTRAL model in its current form can accurately predict the overall spatial quality of automotive audio systems, MUSHRA listening tests using headphone auralisation with head tracking were conducted to collect results to be compared against predictions by the model. Based on guideline criteria, the model in its current form could not accurately predict the overall spatial quality of automotive audio systems. To improve prediction performance, the QESTRAL model was recalibrated and modified using existing metrics of the model, those that were proposed from the literature review, and newly developed metrics. The most important metrics for predicting the overall spatial quality of automotive audio systems included those that were interaural cross-correlation (IACC) based, relate to localisation of the frontal audio scene, and account for the perceived scene width in front of the listener. Modifying the model for automotive audio systems did not invalidate its use for domestic audio systems. The resulting model predicts the overall spatial quality of 2- and 5-channel automotive audio systems with a cross-validation performance of R. 2 = 0.85 and root-mean-squareerror (RMSE) = 11.03%.
Aeronautical audio broadcasting via satellite
NASA Technical Reports Server (NTRS)
Tzeng, Forrest F.
1993-01-01
A system design for aeronautical audio broadcasting, with C-band uplink and L-band downlink, via Inmarsat space segments is presented. Near-transparent-quality compression of 5-kHz bandwidth audio at 20.5 kbit/s is achieved based on a hybrid technique employing linear predictive modeling and transform-domain residual quantization. Concatenated Reed-Solomon/convolutional codes with quadrature phase shift keying are selected for bandwidth and power efficiency. RF bandwidth at 25 kHz per channel, and a decoded bit error rate at 10(exp -6) with E(sub b)/N(sub o) at 3.75 dB are obtained. An interleaver, scrambler, modem synchronization, and frame format were designed, and frequency-division multiple access was selected over code-division multiple access. A link budget computation based on a worst-case scenario indicates sufficient system power margins. Transponder occupancy analysis for 72 audio channels demonstrates ample remaining capacity to accommodate emerging aeronautical services.
Converting laserdisc video to digital video: a demonstration project using brain animations.
Jao, C S; Hier, D B; Brint, S U
1995-01-01
Interactive laserdiscs are of limited value in large group learning situations due to the expense of establishing multiple workstations. The authors implemented an alternative to laserdisc video by using indexed digital video combined with an expert system. High-quality video was captured from a laserdisc player and combined with waveform audio into an audio-video-interleave (AVI) file format in the Microsoft Video-for-Windows environment (Microsoft Corp., Seattle, WA). With the use of an expert system, a knowledge-based computer program provided random access to these indexed AVI files. The program can be played on any multimedia computer without the need for laserdiscs. This system offers a high level of interactive video without the overhead and cost of a laserdisc player.
One size does not fit all: older adults benefit from redundant text in multimedia instruction
Fenesi, Barbara; Vandermorris, Susan; Kim, Joseph A.; Shore, David I.; Heisz, Jennifer J.
2015-01-01
The multimedia design of presentations typically ignores that younger and older adults have varying cognitive strengths and weaknesses. We examined whether differential instructional design may enhance learning in these populations. Younger and older participants viewed one of three computer-based presentations: Audio only (narration), Redundant (audio narration with redundant text), or Complementary (audio narration with non-redundant text and images). Younger participants learned better when audio narration was paired with relevant images compared to when audio narration was paired with redundant text. However, older participants learned best when audio narration was paired with redundant text. Younger adults, who presumably have a higher working memory capacity (WMC), appear to benefit more from complementary information that may drive deeper conceptual processing. In contrast, older adults learn better from presentations that support redundant coding across modalities, which may help mitigate the effects of age-related decline in WMC. Additionally, several misconceptions of design quality appeared across age groups: both younger and older participants positively rated less effective designs. Findings suggest that one-size does not fit all, with older adults requiring unique multimedia design tailored to their cognitive abilities for effective learning. PMID:26284000
One size does not fit all: older adults benefit from redundant text in multimedia instruction.
Fenesi, Barbara; Vandermorris, Susan; Kim, Joseph A; Shore, David I; Heisz, Jennifer J
2015-01-01
The multimedia design of presentations typically ignores that younger and older adults have varying cognitive strengths and weaknesses. We examined whether differential instructional design may enhance learning in these populations. Younger and older participants viewed one of three computer-based presentations: Audio only (narration), Redundant (audio narration with redundant text), or Complementary (audio narration with non-redundant text and images). Younger participants learned better when audio narration was paired with relevant images compared to when audio narration was paired with redundant text. However, older participants learned best when audio narration was paired with redundant text. Younger adults, who presumably have a higher working memory capacity (WMC), appear to benefit more from complementary information that may drive deeper conceptual processing. In contrast, older adults learn better from presentations that support redundant coding across modalities, which may help mitigate the effects of age-related decline in WMC. Additionally, several misconceptions of design quality appeared across age groups: both younger and older participants positively rated less effective designs. Findings suggest that one-size does not fit all, with older adults requiring unique multimedia design tailored to their cognitive abilities for effective learning.
Review of Audio Interfacing Literature for Computer-Assisted Music Instruction.
ERIC Educational Resources Information Center
Watanabe, Nan
1980-01-01
Presents a review of the literature dealing with audio devices used in computer assisted music instruction and discusses the need for research and development of reliable, cost-effective, random access audio hardware. (Author)
ERIC Educational Resources Information Center
Aleman-Centeno, Josefina R.
1983-01-01
Discusses the development and evaluation of CAVIS, which consists of an Apple microcomputer used with audiovisual dialogs. Includes research on the effects of three conditions: (1) computer with audio and visual, (2) computer with audio alone and (3) audio alone in short-term and long-term recall. (EKN)
Value of audio-enhanced handheld computers over paper surveys with adolescents.
Trapl, Erika S; Taylor, H Gerry; Colabianchi, Natalie; Litaker, David; Borawski, Elaine A
2013-01-01
To examine the impact of 3 data collection modes on the number of questions answered, data quality, and student preference. 275 urban seventh-grade students were recruited and randomly assigned to complete a paper survey (SAQ), PDA survey (PDA), or PDA survey with audio (APDA). Students completed a paper debriefing survey. APDA respondents completed significantly more questions compared to SAQ and PDA. PDA and APDA had significantly less missing data than did SAQ. No differences were found for student evaluation. Strong benefits may be gained by the use of APDA for adolescent school-based data collection.
A History of Commitment in CALL.
ERIC Educational Resources Information Center
Jamieson, Joan
The evolution of computer-assisted language learning (CALL) is examined, focusing on what has changed and what has not changed much during that time. A variety of changes are noted: the development of multimedia capabilities, color, animation, and technical improvement of audio and video quality; availability of databases, better fit between…
The Lincoln Training System: A Summary Report.
ERIC Educational Resources Information Center
Butman, Robert C.; Frick, Frederick C.
The current status of the Lincoln Training System (LTS) is reported. This document describes LTS as a computer supported microfiche system which: 1) provides random access to voice quality audio and to graphics; 2) supports student-controlled interactive processes; and 3) functions in a variety of environments. The report offers a detailed…
Reduction in time-to-sleep through EEG based brain state detection and audio stimulation.
Zhuo Zhang; Cuntai Guan; Ti Eu Chan; Juanhong Yu; Aung Aung Phyo Wai; Chuanchu Wang; Haihong Zhang
2015-08-01
We developed an EEG- and audio-based sleep sensing and enhancing system, called iSleep (interactive Sleep enhancement apparatus). The system adopts a closed-loop approach which optimizes the audio recording selection based on user's sleep status detected through our online EEG computing algorithm. The iSleep prototype comprises two major parts: 1) a sleeping mask integrated with a single channel EEG electrode and amplifier, a pair of stereo earphones and a microcontroller with wireless circuit for control and data streaming; 2) a mobile app to receive EEG signals for online sleep monitoring and audio playback control. In this study we attempt to validate our hypothesis that appropriate audio stimulation in relation to brain state can induce faster onset of sleep and improve the quality of a nap. We conduct experiments on 28 healthy subjects, each undergoing two nap sessions - one with a quiet background and one with our audio-stimulation. We compare the time-to-sleep in both sessions between two groups of subjects, e.g., fast and slow sleep onset groups. The p-value obtained from Wilcoxon Signed Rank Test is 1.22e-04 for slow onset group, which demonstrates that iSleep can significantly reduce the time-to-sleep for people with difficulty in falling sleep.
Federal Register 2010, 2011, 2012, 2013, 2014
2010-09-28
... environment. Through the IRIS Program, EPA provides the highest quality science- based human health... external review draft human health assessment titled, ``Toxicological Review of Urea: In Support of Summary... register, please indicate if you will need audio-visual equipment (e.g., laptop computer and slide...
The Use of Audio and Animation in Computer Based Instruction.
ERIC Educational Resources Information Center
Koroghlanian, Carol; Klein, James D.
This study investigated the effects of audio, animation, and spatial ability in a computer-based instructional program for biology. The program presented instructional material via test or audio with lean text and included eight instructional sequences presented either via static illustrations or animations. High school students enrolled in a…
Bertollo, David N; Alexander, Mary Jane; Shinn, Marybeth; Aybar, Jalila B
2007-06-01
This column describes the nonproprietary software Talker, used to adapt screening instruments to audio computer-assisted self-interviewing (ACASI) systems for low-literacy populations and other populations. Talker supports ease of programming, multiple languages, on-site scoring, and the ability to update a central research database. Key features include highly readable text display, audio presentation of questions and audio prompting of answers, and optional touch screen input. The scripting language for adapting instruments is briefly described as well as two studies in which respondents provided positive feedback on its use.
Le, Linh Cu; Vu, Lan T H
2012-10-01
Globally, population surveys on HIV/AIDS and other sensitive topics have been using audio computer-assisted self interview for many years. This interview technique, however, is still new to Vietnam and little is known about its application and impact in general population surveys. One plausible hypothesis is that residents of Vietnam interviewed using this technique may provide a higher response rate and be more willing to reveal their true behaviors than if interviewed with traditional methods. This study aims to compare audio computer-assisted self interview with traditional face-to-face personal interview and self-administered interview with regard to rates of refusal and affirmative responses to questions on sensitive topics related to HIV/AIDS. In June 2010, a randomized study was conducted in three cities (Ha Noi, Da Nan and Can Tho), using a sample of 4049 residents aged 15 to 49 years. Respondents were randomly assigned to one of three interviewing methods: audio computer-assisted self interview, personal face-to-face interview, and self-administered paper interview. Instead of providing answers directly to interviewer questions as with traditional methods, audio computer-assisted self-interview respondents read the questions displayed on a laptop screen, while listening to the questions through audio headphones, then entered responses using a laptop keyboard. A MySQL database was used for data management and SPSS statistical package version 18 used for data analysis with bivariate and multivariate statistical techniques. Rates of high risk behaviors and mean values of continuous variables were compared for the three data collection methods. Audio computer-assisted self interview showed advantages over comparison techniques, achieving lower refusal rates and reporting higher prevalence of some sensitive and risk behaviors (perhaps indication of more truthful answers). Premarital sex was reported by 20.4% in the audio computer-assisted self-interview survey group, versus 11.4% in the face-to-face group and 11.1% in the self-administered paper questionnaire group. The pattern was consistent for both male and female respondents and in both urban and rural settings. Men in the audio computer-assisted self-interview group also reported higher levels of high-risk sexual behavior--such as sex with sex workers and a higher average number of sexual partners--than did women in the same group. Importantly, item refusal rates on sensitive topics tended to be lower with audio computer-assisted self interview than with the other two methods. Combined with existing data from other countries and previous studies in Vietnam, these findings suggest that researchers should consider using audio computer-assisted self interview for future studies of sensitive and stigmatized topics, especially for men.
Benchmarking multimedia performance
NASA Astrophysics Data System (ADS)
Zandi, Ahmad; Sudharsanan, Subramania I.
1998-03-01
With the introduction of faster processors and special instruction sets tailored to multimedia, a number of exciting applications are now feasible on the desktops. Among these is the DVD playback consisting, among other things, of MPEG-2 video and Dolby digital audio or MPEG-2 audio. Other multimedia applications such as video conferencing and speech recognition are also becoming popular on computer systems. In view of this tremendous interest in multimedia, a group of major computer companies have formed, Multimedia Benchmarks Committee as part of Standard Performance Evaluation Corp. to address the performance issues of multimedia applications. The approach is multi-tiered with three tiers of fidelity from minimal to full compliant. In each case the fidelity of the bitstream reconstruction as well as quality of the video or audio output are measured and the system is classified accordingly. At the next step the performance of the system is measured. In many multimedia applications such as the DVD playback the application needs to be run at a specific rate. In this case the measurement of the excess processing power, makes all the difference. All these make a system level, application based, multimedia benchmark very challenging. Several ideas and methodologies for each aspect of the problems will be presented and analyzed.
Federal Register 2010, 2011, 2012, 2013, 2014
2011-08-31
... found in the environment. Through the IRIS Program, EPA provides the highest quality science-based human... for the external review draft human health assessment titled, ``Toxicological Review of n-Butanol: In... will need audio-visual equipment (e.g., laptop computer and slide projector). In general, each...
Sounds of silence: How to animate virtual worlds with sound
NASA Technical Reports Server (NTRS)
Astheimer, Peter
1993-01-01
Sounds are an integral and sometimes annoying part of our daily life. Virtual worlds which imitate natural environments gain a lot of authenticity from fast, high quality visualization combined with sound effects. Sounds help to increase the degree of immersion for human dwellers in imaginary worlds significantly. The virtual reality toolkit of IGD (Institute for Computer Graphics) features a broad range of standard visual and advanced real-time audio components which interpret an object-oriented definition of the scene. The virtual reality system 'Virtual Design' realized with the toolkit enables the designer of virtual worlds to create a true audiovisual environment. Several examples on video demonstrate the usage of the audio features in Virtual Design.
Audiovisual quality evaluation of low-bitrate video
NASA Astrophysics Data System (ADS)
Winkler, Stefan; Faller, Christof
2005-03-01
Audiovisual quality assessment is a relatively unexplored topic. We designed subjective experiments for audio, video, and audiovisual quality using content and encoding parameters representative of video for mobile applications. Our focus were the MPEG-4 AVC (a.k.a. H.264) and AAC coding standards. Our goals in this study are two-fold: we want to understand the interactions between audio and video in terms of perceived audiovisual quality, and we use the subjective data to evaluate the prediction performance of our non-reference video and audio quality metrics.
High-Fidelity Piezoelectric Audio Device
NASA Technical Reports Server (NTRS)
Woodward, Stanley E.; Fox, Robert L.; Bryant, Robert G.
2003-01-01
ModalMax is a very innovative means of harnessing the vibration of a piezoelectric actuator to produce an energy efficient low-profile device with high-bandwidth high-fidelity audio response. The piezoelectric audio device outperforms many commercially available speakers made using speaker cones. The piezoelectric device weighs substantially less (4 g) than the speaker cones which use magnets (10 g). ModalMax devices have extreme fabrication simplicity. The entire audio device is fabricated by lamination. The simplicity of the design lends itself to lower cost. The piezoelectric audio device can be used without its acoustic chambers and thereby resulting in a very low thickness of 0.023 in. (0.58 mm). The piezoelectric audio device can be completely encapsulated, which makes it very attractive for use in wet environments. Encapsulation does not significantly alter the audio response. Its small size (see Figure 1) is applicable to many consumer electronic products, such as pagers, portable radios, headphones, laptop computers, computer monitors, toys, and electronic games. The audio device can also be used in automobile or aircraft sound systems.
Trend Alert: A History Teacher's Guide to Using Podcasts in the Classroom
ERIC Educational Resources Information Center
Swan, Kathleen Owings; Hofer, Mark
2009-01-01
A "podcast" (an amalgam of the word broadcast and the iPod digital audio player) is essentially a broadcast of digital audio files on the web that users can listen to on their computer or digital audio player (e.g., iPod). Podcasts can be automatically delivered to an iPod or computer whenever new content is available. This unique feature of…
Handschu, René; Littmann, Rebekka; Reulbach, Udo; Gaul, Charly; Heckmann, Josef G; Neundörfer, Bernhard; Scibor, Mateusz
2003-12-01
In acute stroke care, rapid but careful evaluation of patients is mandatory but requires an experienced stroke neurologist. Telemedicine offers the possibility of bringing such expertise quickly to more patients. This study tested for the first time whether remote video examination is feasible and reliable when applied in emergency stroke care using the National Institutes of Health Stroke Scale (NIHSS). We used a novel multimedia telesupport system for transfer of real-time video sequences and audio data. The remote examiner could direct the set-top camera and zoom from distant overviews to close-ups from the personal computer in his office. Acute stroke patients admitted to our stroke unit were examined on admission in the emergency room. Standardized examination was performed by use of the NIHSS (German version) via telemedicine and compared with bedside application. In this pilot study, 41 patients were examined. Total examination time was 11.4 minutes on average (range, 8 to 18 minutes). None of the examinations had to be stopped or interrupted for technical reasons, although minor problems (brightness, audio quality) with influence on the examination process occurred in 2 sessions. Unweighted kappa coefficients ranged from 0.44 to 0.89; weighted kappa coefficients, from 0.85 to 0.99. Remote examination of acute stroke patients with a computer-based telesupport system is feasible and reliable when applied in the emergency room; interrater agreement was good to excellent in all items. For more widespread use, some problems that emerge from details like brightness, optimal camera position, and audio quality should be solved.
Audio-Enhanced Computer Assisted Learning and Computer Controlled Audio-Instruction.
ERIC Educational Resources Information Center
Miller, K.; And Others
1983-01-01
Describes aspects of use of a microcomputer linked with a cassette recorder as a peripheral to enhance computer-assisted learning (CAL) and a microcomputer-controlled tape recorder linked with a microfiche reader in a commercially available teaching system. References and a listing of control programs are appended. (EJS)
The Use of Audio in Computer-Based Instruction.
ERIC Educational Resources Information Center
Koroghlanian, Carol M.; Sullivan, Howard J.
This study investigated the effects of audio and text density on the achievement, time-in-program, and attitudes of 134 undergraduates. Data concerning the subjects' preexisting computer skills and experience, as well as demographic information, were also collected. The instruction in visual design principles was delivered by computer and included…
The power of digital audio in interactive instruction: An unexploited medium
DOE Office of Scientific and Technical Information (OSTI.GOV)
Pratt, J.; Trainor, M.
1989-01-01
Widespread use of audio in computer-based training (CBT) occurred with the advent of the interactive videodisc technology. This paper discusses the alternative of digital audio, which, unlike videodisc audio, enables one to rapidly revise the audio used in the CBT and which may be used in nonvideo CBT applications as well. We also discuss techniques used in audio script writing, editing, and production. Results from evaluations indicate a high degree of user satisfaction. 4 refs.
Method for Reading Sensors and Controlling Actuators Using Audio Interfaces of Mobile Devices
Aroca, Rafael V.; Burlamaqui, Aquiles F.; Gonçalves, Luiz M. G.
2012-01-01
This article presents a novel closed loop control architecture based on audio channels of several types of computing devices, such as mobile phones and tablet computers, but not restricted to them. The communication is based on an audio interface that relies on the exchange of audio tones, allowing sensors to be read and actuators to be controlled. As an application example, the presented technique is used to build a low cost mobile robot, but the system can also be used in a variety of mechatronics applications and sensor networks, where smartphones are the basic building blocks. PMID:22438726
Method for reading sensors and controlling actuators using audio interfaces of mobile devices.
Aroca, Rafael V; Burlamaqui, Aquiles F; Gonçalves, Luiz M G
2012-01-01
This article presents a novel closed loop control architecture based on audio channels of several types of computing devices, such as mobile phones and tablet computers, but not restricted to them. The communication is based on an audio interface that relies on the exchange of audio tones, allowing sensors to be read and actuators to be controlled. As an application example, the presented technique is used to build a low cost mobile robot, but the system can also be used in a variety of mechatronics applications and sensor networks, where smartphones are the basic building blocks.
Subjective evaluation and electroacoustic theoretical validation of a new approach to audio upmixing
NASA Astrophysics Data System (ADS)
Usher, John S.
Audio signal processing systems for converting two-channel (stereo) recordings to four or five channels are increasingly relevant. These audio upmixers can be used with conventional stereo sound recordings and reproduced with multichannel home theatre or automotive loudspeaker audio systems to create a more engaging and natural-sounding listening experience. This dissertation discusses existing approaches to audio upmixing for recordings of musical performances and presents specific design criteria for a system to enhance spatial sound quality. A new upmixing system is proposed and evaluated according to these criteria and a theoretical model for its behavior is validated using empirical measurements. The new system removes short-term correlated components from two electronic audio signals using a pair of adaptive filters, updated according to a frequency domain implementation of the normalized-least-means-square algorithm. The major difference of the new system with all extant audio upmixers is that unsupervised time-alignment of the input signals (typically, by up to +/-10 ms) as a function of frequency (typically, using a 1024-band equalizer) is accomplished due to the non-minimum phase adaptive filter. Two new signals are created from the weighted difference of the inputs, and are then radiated with two loudspeakers behind the listener. According to the consensus in the literature on the effect of interaural correlation on auditory image formation, the self-orthogonalizing properties of the algorithm ensure minimal distortion of the frontal source imagery and natural-sounding, enveloping reverberance (ambiance) imagery. Performance evaluation of the new upmix system was accomplished in two ways: Firstly, using empirical electroacoustic measurements which validate a theoretical model of the system; and secondly, with formal listening tests which investigated auditory spatial imagery with a graphical mapping tool and a preference experiment. Both electroacoustic and subjective methods investigated system performance with a variety of test stimuli for solo musical performances reproduced using a loudspeaker in an orchestral concert-hall and recorded using different microphone techniques. The objective and subjective evaluations combined with a comparative study with two commercial systems demonstrate that the proposed system provides a new, computationally practical, high sound quality solution to upmixing.
NASA Astrophysics Data System (ADS)
Barbieri, Ivano; Lambruschini, Paolo; Raggio, Marco; Stagnaro, Riccardo
2007-12-01
The increase in the availability of bandwidth for wireless links, network integration, and the computational power on fixed and mobile platforms at affordable costs allows nowadays for the handling of audio and video data, their quality making them suitable for medical application. These information streams can support both continuous monitoring and emergency situations. According to this scenario, the authors have developed and implemented the mobile communication system which is described in this paper. The system is based on ITU-T H.323 multimedia terminal recommendation, suitable for real-time data/video/audio and telemedical applications. The audio and video codecs, respectively, H.264 and G723.1, were implemented and optimized in order to obtain high performance on the system target processors. Offline media streaming storage and retrieval functionalities were supported by integrating a relational database in the hospital central system. The system is based on low-cost consumer technologies such as general packet radio service (GPRS) and wireless local area network (WLAN or WiFi) for lowband data/video transmission. Implementation and testing were carried out for medical emergency and telemedicine application. In this paper, the emergency case study is described.
Direct broadcast satellite-audio, portable and mobile reception tradeoffs
NASA Technical Reports Server (NTRS)
Golshan, Nasser
1992-01-01
This paper reports on the findings of a systems tradeoffs study on direct broadcast satellite-radio (DBS-R). Based on emerging advanced subband and transform audio coding systems, four ranges of bit rates: 16-32 kbps, 48-64 kbps, 96-128 kbps and 196-256 kbps are identified for DBS-R. The corresponding grades of audio quality will be subjectively comparable to AM broadcasting, monophonic FM, stereophonic FM, and CD quality audio, respectively. The satellite EIRP's needed for mobile DBS-R reception in suburban areas are sufficient for portable reception in most single family houses when allowance is made for the higher G/T of portable table-top receivers. As an example, the variation of the space segment cost as a function of frequency, audio quality, coverage capacity, and beam size is explored for a typical DBS-R system.
Digital Multicasting of Multiple Audio Streams
NASA Technical Reports Server (NTRS)
Macha, Mitchell; Bullock, John
2007-01-01
The Mission Control Center Voice Over Internet Protocol (MCC VOIP) system (see figure) comprises hardware and software that effect simultaneous, nearly real-time transmission of as many as 14 different audio streams to authorized listeners via the MCC intranet and/or the Internet. The original version of the MCC VOIP system was conceived to enable flight-support personnel located in offices outside a spacecraft mission control center to monitor audio loops within the mission control center. Different versions of the MCC VOIP system could be used for a variety of public and commercial purposes - for example, to enable members of the general public to monitor one or more NASA audio streams through their home computers, to enable air-traffic supervisors to monitor communication between airline pilots and air-traffic controllers in training, and to monitor conferences among brokers in a stock exchange. At the transmitting end, the audio-distribution process begins with feeding the audio signals to analog-to-digital converters. The resulting digital streams are sent through the MCC intranet, using a user datagram protocol (UDP), to a server that converts them to encrypted data packets. The encrypted data packets are then routed to the personal computers of authorized users by use of multicasting techniques. The total data-processing load on the portion of the system upstream of and including the encryption server is the total load imposed by all of the audio streams being encoded, regardless of the number of the listeners or the number of streams being monitored concurrently by the listeners. The personal computer of a user authorized to listen is equipped with special- purpose MCC audio-player software. When the user launches the program, the user is prompted to provide identification and a password. In one of two access- control provisions, the program is hard-coded to validate the user s identity and password against a list maintained on a domain-controller computer at the MCC. In the other access-control provision, the program verifies that the user is authorized to have access to the audio streams. Once both access-control checks are completed, the audio software presents a graphical display that includes audiostream-selection buttons and volume-control sliders. The user can select all or any subset of the available audio streams and can adjust the volume of each stream independently of that of the other streams. The audio-player program spawns a "read" process for the selected stream(s). The spawned process sends, to the router(s), a "multicast-join" request for the selected streams. The router(s) responds to the request by sending the encrypted multicast packets to the spawned process. The spawned process receives the encrypted multicast packets and sends a decryption packet to audio-driver software. As the volume or muting features are changed by the user, interrupts are sent to the spawned process to change the corresponding attributes sent to the audio-driver software. The total latency of this system - that is, the total time from the origination of the audio signals to generation of sound at a listener s computer - lies between four and six seconds.
Watermarking 3D Objects for Verification
1999-01-01
signal (audio/ image /video) pro- cessing and steganography fields, and even newer to the computer graphics community. Inherently, digital watermarking of...quality images , and digital video. The field of digital watermarking is relatively new, and many of its terms have not been well defined. Among the dif...ferent media types, watermarking of 2D still images is comparatively better studied. Inherently, digital water- marking of 3D objects remains a
Audio-Tutorial Instruction in Medicine.
ERIC Educational Resources Information Center
Boyle, Gloria J.; Herrick, Merlyn C.
This progress report concerns an audio-tutorial approach used at the University of Missouri-Columbia School of Medicine. Instructional techniques such as slide-tape presentations, compressed speech audio tapes, computer-assisted instruction (CAI), motion pictures, television, microfiche, and graphic and printed materials have been implemented,…
Effect of tape recording on perturbation measures.
Jiang, J; Lin, E; Hanson, D G
1998-10-01
Tape recorders have been shown to affect measures of voice perturbation. Few studies, however, have been conducted to quantitatively justify the use or exclusion of certain types of recorders in voice perturbation studies. This study used sinusoidal and triangular waves and synthesized vowels to compare perturbation measures extracted from directly digitized signals with those recorded and played back through various tape recorders, including 3 models of digital audio tape recorders, 2 models of analog audio cassette tape recorders, and 2 models of video tape recorders. Signal contamination for frequency perturbation values was found to be consistently minimal with digital recorders (percent jitter = 0.01%-0.02%), mildly increased with video recorders (0.05%-0.10%), moderately increased with a high-quality analog audio cassette tape recorder (0.15%), and most prominent with a low-quality analog audio cassette tape recorder (0.24%). Recorder effect on amplitude perturbation measures was lowest in digital recorders (percent shimmer = 0.09%-0.20%), mildly to moderately increased in video recorders and a high-quality analog audio cassette tape recorder (0.25%-0.45%), and most prominent in a low-quality analog audio cassette tape recorder (0.98%). The effect of cassette tape material, length of spooled tape, and duration of analysis were also tested and are discussed.
Audio-visual affective expression recognition
NASA Astrophysics Data System (ADS)
Huang, Thomas S.; Zeng, Zhihong
2007-11-01
Automatic affective expression recognition has attracted more and more attention of researchers from different disciplines, which will significantly contribute to a new paradigm for human computer interaction (affect-sensitive interfaces, socially intelligent environments) and advance the research in the affect-related fields including psychology, psychiatry, and education. Multimodal information integration is a process that enables human to assess affective states robustly and flexibly. In order to understand the richness and subtleness of human emotion behavior, the computer should be able to integrate information from multiple sensors. We introduce in this paper our efforts toward machine understanding of audio-visual affective behavior, based on both deliberate and spontaneous displays. Some promising methods are presented to integrate information from both audio and visual modalities. Our experiments show the advantage of audio-visual fusion in affective expression recognition over audio-only or visual-only approaches.
Audio Frequency Analysis in Mobile Phones
ERIC Educational Resources Information Center
Aguilar, Horacio Munguía
2016-01-01
A new experiment using mobile phones is proposed in which its audio frequency response is analyzed using the audio port for inputting external signal and getting a measurable output. This experiment shows how the limited audio bandwidth used in mobile telephony is the main cause of the poor speech quality in this service. A brief discussion is…
Babjack, Destiny L; Cernicky, Brandon; Sobotka, Andrew J; Basler, Lee; Struthers, Devon; Kisic, Richard; Barone, Kimberly; Zuccolotto, Anthony P
2015-09-01
Using differing computer platforms and audio output devices to deliver audio stimuli often introduces (1) substantial variability across labs and (2) variable time between the intended and actual sound delivery (the sound onset latency). Fast, accurate audio onset latencies are particularly important when audio stimuli need to be delivered precisely as part of studies that depend on accurate timing (e.g., electroencephalographic, event-related potential, or multimodal studies), or in multisite studies in which standardization and strict control over the computer platforms used is not feasible. This research describes the variability introduced by using differing configurations and introduces a novel approach to minimizing audio sound latency and variability. A stimulus presentation and latency assessment approach is presented using E-Prime and Chronos (a new multifunction, USB-based data presentation and collection device). The present approach reliably delivers audio stimuli with low latencies that vary by ≤1 ms, independent of hardware and Windows operating system (OS)/driver combinations. The Chronos audio subsystem adopts a buffering, aborting, querying, and remixing approach to the delivery of audio, to achieve a consistent 1-ms sound onset latency for single-sound delivery, and precise delivery of multiple sounds that achieves standard deviations of 1/10th of a millisecond without the use of advanced scripting. Chronos's sound onset latencies are small, reliable, and consistent across systems. Testing of standard audio delivery devices and configurations highlights the need for careful attention to consistency between labs, experiments, and multiple study sites in their hardware choices, OS selections, and adoption of audio delivery systems designed to sidestep the audio latency variability issue.
Low-delay predictive audio coding for the HIVITS HDTV codec
NASA Astrophysics Data System (ADS)
McParland, A. K.; Gilchrist, N. H. C.
1995-01-01
The status of work relating to predictive audio coding, as part of the European project on High Quality Video Telephone and HD(TV) Systems (HIVITS), is reported. The predictive coding algorithm is developed, along with six-channel audio coding and decoding hardware. Demonstrations of the audio codec operating in conjunction with the video codec, are given.
Tune in the Net with RealAudio.
ERIC Educational Resources Information Center
Buchanan, Larry
1997-01-01
Describes how to connect to the RealAudio Web site to download a player that provides sound from Web pages to the computer through streaming technology. Explains hardware and software requirements and provides addresses for other RealAudio Web sites are provided, including weather information and current news. (LRW)
Three-Dimensional Audio Client Library
NASA Technical Reports Server (NTRS)
Rizzi, Stephen A.
2005-01-01
The Three-Dimensional Audio Client Library (3DAudio library) is a group of software routines written to facilitate development of both stand-alone (audio only) and immersive virtual-reality application programs that utilize three-dimensional audio displays. The library is intended to enable the development of three-dimensional audio client application programs by use of a code base common to multiple audio server computers. The 3DAudio library calls vendor-specific audio client libraries and currently supports the AuSIM Gold-Server and Lake Huron audio servers. 3DAudio library routines contain common functions for (1) initiation and termination of a client/audio server session, (2) configuration-file input, (3) positioning functions, (4) coordinate transformations, (5) audio transport functions, (6) rendering functions, (7) debugging functions, and (8) event-list-sequencing functions. The 3DAudio software is written in the C++ programming language and currently operates under the Linux, IRIX, and Windows operating systems.
Federal Register 2010, 2011, 2012, 2013, 2014
2010-02-23
... the essential bus. The disabled equipment could include the autopilot, anti-skid system, hydraulic indicator, spoiler system, pilot primary flight display, audio panel, or the 1 air data computer. This... system, pilot primary flight display, audio panel, or the 1 air data computer. This failure could lead to...
ERIC Educational Resources Information Center
Dow, James
Ways in which computers and audio tape recorder techniques were used to record, index, and present data collected during two summers of field work in a rural area of Mexico are described. The research goal was to study the Otomi Indian shamans. Two computers were used: the Honeywell 6800 DPS-2 and the Osborne-1 microcomputer. The database system…
Informed spectral analysis: audio signal parameter estimation using side information
NASA Astrophysics Data System (ADS)
Fourer, Dominique; Marchand, Sylvain
2013-12-01
Parametric models are of great interest for representing and manipulating sounds. However, the quality of the resulting signals depends on the precision of the parameters. When the signals are available, these parameters can be estimated, but the presence of noise decreases the resulting precision of the estimation. Furthermore, the Cramér-Rao bound shows the minimal error reachable with the best estimator, which can be insufficient for demanding applications. These limitations can be overcome by using the coding approach which consists in directly transmitting the parameters with the best precision using the minimal bitrate. However, this approach does not take advantage of the information provided by the estimation from the signal and may require a larger bitrate and a loss of compatibility with existing file formats. The purpose of this article is to propose a compromised approach, called the 'informed approach,' which combines analysis with (coded) side information in order to increase the precision of parameter estimation using a lower bitrate than pure coding approaches, the audio signal being known. Thus, the analysis problem is presented in a coder/decoder configuration where the side information is computed and inaudibly embedded into the mixture signal at the coder. At the decoder, the extra information is extracted and is used to assist the analysis process. This study proposes applying this approach to audio spectral analysis using sinusoidal modeling which is a well-known model with practical applications and where theoretical bounds have been calculated. This work aims at uncovering new approaches for audio quality-based applications. It provides a solution for challenging problems like active listening of music, source separation, and realistic sound transformations.
Video conference quality assessment based on cooperative sensing of video and audio
NASA Astrophysics Data System (ADS)
Wang, Junxi; Chen, Jialin; Tian, Xin; Zhou, Cheng; Zhou, Zheng; Ye, Lu
2015-12-01
This paper presents a method to video conference quality assessment, which is based on cooperative sensing of video and audio. In this method, a proposed video quality evaluation method is used to assess the video frame quality. The video frame is divided into noise image and filtered image by the bilateral filters. It is similar to the characteristic of human visual, which could also be seen as a low-pass filtering. The audio frames are evaluated by the PEAQ algorithm. The two results are integrated to evaluate the video conference quality. A video conference database is built to test the performance of the proposed method. It could be found that the objective results correlate well with MOS. Then we can conclude that the proposed method is efficiency in assessing video conference quality.
ERIC Educational Resources Information Center
Ludlow, Barbara L.; Foshay, John B.; Duff, Michael C.
Video presentations of teaching episodes in home, school, and community settings and audio recordings of parents' and professionals' views can be important adjuncts to personnel preparation in special education. This paper describes instructional applications of digital media and outlines steps in producing audio and video segments. Digital audio…
The Effect of Audio and Animation in Multimedia Instruction
ERIC Educational Resources Information Center
Koroghlanian, Carol; Klein, James D.
2004-01-01
This study investigated the effects of audio, animation, and spatial ability in a multimedia computer program for high school biology. Participants completed a multimedia program that presented content by way of text or audio with lean text. In addition, several instructional sequences were presented either with static illustrations or animations.…
Wofford, J L; Currin, D; Michielutte, R; Wofford, M M
2001-04-20
Inadequate reading literacy is a major barrier to better educating patients. Despite its high prevalence, practical solutions for detecting and overcoming low literacy in a busy clinical setting remain elusive. In exploring the potential role for the multimedia computer in improving office-based patient education, we compared the accuracy of information captured from audio-computer interviewing of patients with that obtained from subsequent verbal questioning. Adult medicine clinic, urban community health center Convenience sample of patients awaiting clinic appointments (n = 59). Exclusion criteria included obvious psychoneurologic impairment or primary language other than English. A multimedia computer presentation that used audio-computer interviewing with localized imagery and voices to elicit responses to 4 questions on prior computer use and cancer risk perceptions. Three patients refused or were unable to interact with the computer at all, and 3 patients required restarting the presentation from the beginning but ultimately completed the computerized survey. Of the 51 evaluable patients (72.5% African-American, 66.7% female, mean age 47.5 [+/- 18.1]), the mean time in the computer presentation was significantly longer with older age and with no prior computer use but did not differ by gender or race. Despite a high proportion of no prior computer use (60.8%), there was a high rate of agreement (88.7% overall) between audio-computer interviewing and subsequent verbal questioning. Audio-computer interviewing is feasible in this urban community health center. The computer offers a partial solution for overcoming literacy barriers inherent in written patient education materials and provides an efficient means of data collection that can be used to better target patients' educational needs.
The Effect of Gloss Type and Mode on Iranian EFL Learners' Reading Comprehension
ERIC Educational Resources Information Center
Sadeghi, Karim; Ahmadi, Negar
2012-01-01
This study investigated the effects of three kinds of gloss conditions that is traditional non-CALL marginal gloss, computer-based audio gloss, and computer-based extended audio gloss, on reading comprehension of Iranian EFL learners. To this end, three experimental and one control groups, each comprising 15 participants, took part in this study.…
Federal Register 2010, 2011, 2012, 2013, 2014
2010-05-17
... system, pilot primary flight display, audio panel, or the 1 air data computer. This failure could lead to... include the autopilot, anti-skid system, hydraulic indicator, spoiler system, pilot primary flight display, audio panel, or the 1 air data computer. This failure could lead to a significant increase in pilot...
ERIC Educational Resources Information Center
Ko, Chao-Jung
2012-01-01
This study investigated the possibility that initial-level learners may acquire oral skills through synchronous computer-mediated communication (SCMC). Twelve Taiwanese French as a foreign language (FFL) students, divided into three groups, were required to conduct a variety of tasks in one of the three learning environments (video/audio, audio,…
LiveDescribe: Can Amateur Describers Create High-Quality Audio Description?
ERIC Educational Resources Information Center
Branje, Carmen J.; Fels, Deborah I.
2012-01-01
Introduction: The study presented here evaluated the usability of the audio description software LiveDescribe and explored the acceptance rates of audio description created by amateur describers who used LiveDescribe to facilitate the creation of their descriptions. Methods: Twelve amateur describers with little or no previous experience with…
Audio Design: Creating Multi-sensory Images for the Mind.
ERIC Educational Resources Information Center
Ferrington, Gary
1994-01-01
Explores the concept of "theater of the mind" and discusses design factors in creating audio works that effectively stimulate mental pictures, including: narrative format in audio scripting; qualities of voice; use of concrete language; music; noise versus silence; and the creation of the illusion of space using monaural, stereophonic,…
Using Text-to-Speech (TTS) for Audio Computer-Assisted Self-Interviewing (ACASI)
ERIC Educational Resources Information Center
Couper, Mick P.; Berglund, Patricia; Kirgis, Nicole; Buageila, Sarrah
2016-01-01
We evaluate the use of text-to-speech (TTS) technology for audio computer-assisted self-interviewing (ACASI). We use a quasi-experimental design, comparing the use of recorded human voice in the 2006-2010 National Survey of Family Growth with the use of TTS in the first year of the 2011-2013 survey, where the essential survey conditions are…
Subtlenoise: sonification of distributed computing operations
NASA Astrophysics Data System (ADS)
Love, P. A.
2015-12-01
The operation of distributed computing systems requires comprehensive monitoring to ensure reliability and robustness. There are two components found in most monitoring systems: one being visually rich time-series graphs and another being notification systems for alerting operators under certain pre-defined conditions. In this paper the sonification of monitoring messages is explored using an architecture that fits easily within existing infrastructures based on mature opensource technologies such as ZeroMQ, Logstash, and Supercollider (a synth engine). Message attributes are mapped onto audio attributes based on broad classification of the message (continuous or discrete metrics) but keeping the audio stream subtle in nature. The benefits of audio rendering are described in the context of distributed computing operations and may provide a less intrusive way to understand the operational health of these systems.
Digital Audio Radio Broadcast Systems Laboratory Testing Nearly Complete
NASA Technical Reports Server (NTRS)
2005-01-01
Radio history continues to be made at the NASA Lewis Research Center with the completion of phase one of the digital audio radio (DAR) testing conducted by the Consumer Electronics Group of the Electronic Industries Association. This satellite, satellite/terrestrial, and terrestrial digital technology will open up new audio broadcasting opportunities both domestically and worldwide. It will significantly improve the current quality of amplitude-modulated/frequency-modulated (AM/FM) radio with a new digitally modulated radio signal and will introduce true compact-disc-quality (CD-quality) sound for the first time. Lewis is hosting the laboratory testing of seven proposed digital audio radio systems and modes. Two of the proposed systems operate in two modes each, making a total of nine systems being tested. The nine systems are divided into the following types of transmission: in-band on-channel (IBOC), in-band adjacent-channel (IBAC), and new bands. The laboratory testing was conducted by the Consumer Electronics Group of the Electronic Industries Association. Subjective assessments of the audio recordings for each of the nine systems was conducted by the Communications Research Center in Ottawa, Canada, under contract to the Electronic Industries Association. The Communications Research Center has the only CCIR-qualified (Consultative Committee for International Radio) audio testing facility in North America. The main goals of the U.S. testing process are to (1) provide technical data to the Federal Communication Commission (FCC) so that it can establish a standard for digital audio receivers and transmitters and (2) provide the receiver and transmitter industries with the proper standards upon which to build their equipment. In addition, the data will be forwarded to the International Telecommunications Union to help in the establishment of international standards for digital audio receivers and transmitters, thus allowing U.S. manufacturers to compete in the world market.
Guidelines for the Production of Audio Materials for Print Handicapped Readers.
ERIC Educational Resources Information Center
National Library of Australia, Canberra.
Procedural guidelines developed by the Audio Standards Committee of the National Library of Australia to help improve the overall quality of production of audio materials for visually handicapped readers are presented. This report covers the following areas: selection of narrators and the narration itself; copyright; recording of books, magazines,…
A compact electroencephalogram recording device with integrated audio stimulation system.
Paukkunen, Antti K O; Kurttio, Anttu A; Leminen, Miika M; Sepponen, Raimo E
2010-06-01
A compact (96 x 128 x 32 mm(3), 374 g), battery-powered, eight-channel electroencephalogram recording device with an integrated audio stimulation system and a wireless interface is presented. The recording device is capable of producing high-quality data, while the operating time is also reasonable for evoked potential studies. The effective measurement resolution is about 4 nV at 200 Hz sample rate, typical noise level is below 0.7 microV(rms) at 0.16-70 Hz, and the estimated operating time is 1.5 h. An embedded audio decoder circuit reads and plays wave sound files stored on a memory card. The activities are controlled by an 8 bit main control unit which allows accurate timing of the stimuli. The interstimulus interval jitter measured is less than 1 ms. Wireless communication is made through bluetooth and the data recorded are transmitted to an external personal computer (PC) interface in real time. The PC interface is implemented with LABVIEW and in addition to data acquisition it also allows online signal processing, data storage, and control of measurement activities such as contact impedance measurement, for example. The practical application of the device is demonstrated in mismatch negativity experiment with three test subjects.
A compact electroencephalogram recording device with integrated audio stimulation system
NASA Astrophysics Data System (ADS)
Paukkunen, Antti K. O.; Kurttio, Anttu A.; Leminen, Miika M.; Sepponen, Raimo E.
2010-06-01
A compact (96×128×32 mm3, 374 g), battery-powered, eight-channel electroencephalogram recording device with an integrated audio stimulation system and a wireless interface is presented. The recording device is capable of producing high-quality data, while the operating time is also reasonable for evoked potential studies. The effective measurement resolution is about 4 nV at 200 Hz sample rate, typical noise level is below 0.7 μVrms at 0.16-70 Hz, and the estimated operating time is 1.5 h. An embedded audio decoder circuit reads and plays wave sound files stored on a memory card. The activities are controlled by an 8 bit main control unit which allows accurate timing of the stimuli. The interstimulus interval jitter measured is less than 1 ms. Wireless communication is made through bluetooth and the data recorded are transmitted to an external personal computer (PC) interface in real time. The PC interface is implemented with LABVIEW® and in addition to data acquisition it also allows online signal processing, data storage, and control of measurement activities such as contact impedance measurement, for example. The practical application of the device is demonstrated in mismatch negativity experiment with three test subjects.
The relationship between basic audio quality and overall listening experience.
Schoeffler, Michael; Herre, Jürgen
2016-09-01
Basic audio quality (BAQ) is a well-known perceptual attribute, which is rated in various listening test methods to measure the performance of audio systems. Unfortunately, when it comes to purchasing audio systems, BAQ might not have a significant influence on the customers' buying decisions since other factors, like brand loyalty, might be more important. In contrast to BAQ, overall listening experience (OLE) is an affective attribute which incorporates all aspects that are important to an individual assessor, including his or her preference for music genre and audio quality. In this work, the relationship between BAQ and OLE is investigated in more detail. To this end, an experiment was carried out, in which participants rated the BAQ and the OLE of music excerpts with different timbral and spatial degradations. In a between-group-design procedure, participants were assigned into two groups, in each of which a different set of stimuli was rated. The results indicate that rating of both attributes, BAQ and OLE, leads to similar rankings, even if a different set of stimuli is rated. In contrast to the BAQ ratings, which were more influenced by timbral than spatial degradations, the OLE ratings were almost equally influenced by timbral and spatial degradations.
A Study of Quality of Service Communication for High-Speed Packet-Switching Computer Sub-Networks
NASA Technical Reports Server (NTRS)
Cui, Zhenqian
1999-01-01
With the development of high-speed networking technology, computer networks, including local-area networks (LANs), wide-area networks (WANs) and the Internet, are extending their traditional roles of carrying computer data. They are being used for Internet telephony, multimedia applications such as conferencing and video on demand, distributed simulations, and other real-time applications. LANs are even used for distributed real-time process control and computing as a cost-effective approach. Differing from traditional data transfer, these new classes of high-speed network applications (video, audio, real-time process control, and others) are delay sensitive. The usefulness of data depends not only on the correctness of received data, but also the time that data are received. In other words, these new classes of applications require networks to provide guaranteed services or quality of service (QoS). Quality of service can be defined by a set of parameters and reflects a user's expectation about the underlying network's behavior. Traditionally, distinct services are provided by different kinds of networks. Voice services are provided by telephone networks, video services are provided by cable networks, and data transfer services are provided by computer networks. A single network providing different services is called an integrated-services network.
Sounding ruins: reflections on the production of an 'audio drift'.
Gallagher, Michael
2015-07-01
This article is about the use of audio media in researching places, which I term 'audio geography'. The article narrates some episodes from the production of an 'audio drift', an experimental environmental sound work designed to be listened to on a portable MP3 player whilst walking in a ruinous landscape. Reflecting on how this work functions, I argue that, as well as representing places, audio geography can shape listeners' attention and bodily movements, thereby reworking places, albeit temporarily. I suggest that audio geography is particularly apt for amplifying the haunted and uncanny qualities of places. I discuss some of the issues raised for research ethics, epistemology and spectral geographies.
Sounding ruins: reflections on the production of an ‘audio drift’
Gallagher, Michael
2014-01-01
This article is about the use of audio media in researching places, which I term ‘audio geography’. The article narrates some episodes from the production of an ‘audio drift’, an experimental environmental sound work designed to be listened to on a portable MP3 player whilst walking in a ruinous landscape. Reflecting on how this work functions, I argue that, as well as representing places, audio geography can shape listeners’ attention and bodily movements, thereby reworking places, albeit temporarily. I suggest that audio geography is particularly apt for amplifying the haunted and uncanny qualities of places. I discuss some of the issues raised for research ethics, epistemology and spectral geographies. PMID:29708107
Fall Detection Using Smartphone Audio Features.
Cheffena, Michael
2016-07-01
An automated fall detection system based on smartphone audio features is developed. The spectrogram, mel frequency cepstral coefficents (MFCCs), linear predictive coding (LPC), and matching pursuit (MP) features of different fall and no-fall sound events are extracted from experimental data. Based on the extracted audio features, four different machine learning classifiers: k-nearest neighbor classifier (k-NN), support vector machine (SVM), least squares method (LSM), and artificial neural network (ANN) are investigated for distinguishing between fall and no-fall events. For each audio feature, the performance of each classifier in terms of sensitivity, specificity, accuracy, and computational complexity is evaluated. The best performance is achieved using spectrogram features with ANN classifier with sensitivity, specificity, and accuracy all above 98%. The classifier also has acceptable computational requirement for training and testing. The system is applicable in home environments where the phone is placed in the vicinity of the user.
ERIC Educational Resources Information Center
Kim, Young-Suk Grace
2016-01-01
Purpose: The primary aim of the present study was to examine whether different ways of presenting narrative stimuli (i.e., live narrative stimuli versus audio-recorded narrative stimuli) influence children's performances on narrative comprehension and oral-retell quality. Method: Children in kindergarten (n = 54), second grade (n = 74), and fourth…
ERIC Educational Resources Information Center
Kim, Young-Suk Grace
2016-01-01
Purpose: The primary aim of the present study was to examine whether different ways of presenting narrative stimuli (i.e., live narrative stimuli versus audio-recorded narrative stimuli) influence children's performances on narrative comprehension and oral-retell quality. Method: Children in kindergarten (n = 54), second grade (n = 74), and fourth…
ERIC Educational Resources Information Center
Wang, Pei-Yu; Huang, Chung-Kai
2015-01-01
This study aims to explore the impact of learner grade, visual cueing, and control design on children's reading achievement of audio e-books with tablet computers. This research was a three-way factorial design where the first factor was learner grade (grade four and six), the second factor was e-book visual cueing (word-based, line-based, and…
A review of lossless audio compression standards and algorithms
NASA Astrophysics Data System (ADS)
Muin, Fathiah Abdul; Gunawan, Teddy Surya; Kartiwi, Mira; Elsheikh, Elsheikh M. A.
2017-09-01
Over the years, lossless audio compression has gained popularity as researchers and businesses has become more aware of the need for better quality and higher storage demand. This paper will analyse various lossless audio coding algorithm and standards that are used and available in the market focusing on Linear Predictive Coding (LPC) specifically due to its popularity and robustness in audio compression, nevertheless other prediction methods are compared to verify this. Advanced representation of LPC such as LSP decomposition techniques are also discussed within this paper.
Mears, Molly; Coonrod, Dean V; Bay, R Curtis; Mills, Terry E; Watkins, Michelle C
2005-09-01
To compare endorsement rates obtained with audio computer-assisted self-interview versus routine prenatal history. A crosssectional study compared items captured with the routine history to those captured with a computer interview (computer screen displaying and computer audio reading questions, with responses entered by touch screen). The subjects were women (n=174) presenting to a public hospital clinic for prenatal care. The prevalence of positive responses using the computer interview was significantly greater (p < 0.01) than with the routine history for induced abortion (16.8% versus 4.0%), lifetime smoking (12.8% versus 5.2%), intimate partner violence (10.0% versus 2.4%), ectopic pregnancy (5.2% versus 1.1%) and family history of mental retardation (6.7% versus 0.6%). Significant differences were not found for history of spontaneous abortion, hypertension, epilepsy, thyroid disease, smoking during pregnancy, gynecologic surgery, abnormal Pap test, neural tube defect or cystic fibrosis family history. However, in all cases, prevalence was equal or greater with the computer interview. Women were more likely to report sensitive and high-risk behavior, such as smoking history, intimate partner violence and elective abortion, with the computer interview. The computer interview displayed equal or increased patient reporting of positive responses and may therefore be an accurate method of obtaining an initial history.
Hallum-Montes, Rachel; Senter, Lindsay; D'Souza, Rohan; Gates-Ferris, Kathryn; Hurlbert, Marc; Anastario, Michael
2014-01-01
This study compares rates of completion of client intake forms (CIFs) collected via three interview modes: audio computer-assisted self-interview (ACASI), face-to-face interview (FFI), and self-administered paper-based interview (SAPI). A total of 303 clients served through the Avon Breast Health Outreach Program (BHOP) were sampled from three U.S. sites. Clients were randomly assigned to complete a standard CIF via one of the three interview modes. Logistic regression analyses demonstrated that clients were significantly more likely to complete the entire CIF via ACASI than either FFI or SAPI. The greatest observed differences were between ACASI and SAPI; clients were almost six times more likely to complete the CIF via ACASI as opposed to SAPI (AOR = 5.8, p < .001). We recommend that where feasible, ACASI be utilized as an effective means of collecting client-level data in healthcare settings. Adoption of ACASI in health centers may translate into higher completion rates of intake forms by clients, as well as reduced burden on clinic staff to enter data and review intake forms for completion. © 2013 National Association for Healthcare Quality.
A centralized audio presentation manager
DOE Office of Scientific and Technical Information (OSTI.GOV)
Papp, A.L. III; Blattner, M.M.
1994-05-16
The centralized audio presentation manager addresses the problems which occur when multiple programs running simultaneously attempt to use the audio output of a computer system. Time dependence of sound means that certain auditory messages must be scheduled simultaneously, which can lead to perceptual problems due to psychoacoustic phenomena. Furthermore, the combination of speech and nonspeech audio is examined; each presents its own problems of perceptibility in an acoustic environment composed of multiple auditory streams. The centralized audio presentation manager receives abstract parameterized message requests from the currently running programs, and attempts to create and present a sonic representation in themore » most perceptible manner through the use of a theoretically and empirically designed rule set.« less
Robot Command Interface Using an Audio-Visual Speech Recognition System
NASA Astrophysics Data System (ADS)
Ceballos, Alexánder; Gómez, Juan; Prieto, Flavio; Redarce, Tanneguy
In recent years audio-visual speech recognition has emerged as an active field of research thanks to advances in pattern recognition, signal processing and machine vision. Its ultimate goal is to allow human-computer communication using voice, taking into account the visual information contained in the audio-visual speech signal. This document presents a command's automatic recognition system using audio-visual information. The system is expected to control the laparoscopic robot da Vinci. The audio signal is treated using the Mel Frequency Cepstral Coefficients parametrization method. Besides, features based on the points that define the mouth's outer contour according to the MPEG-4 standard are used in order to extract the visual speech information.
NASA Astrophysics Data System (ADS)
Nasution, A. B.; Efendi, S.; Suwilo, S.
2018-04-01
The amount of data inserted in the form of audio samples that use 8 bits with LSB algorithm, affect the value of PSNR which resulted in changes in image quality of the insertion (fidelity). So in this research will be inserted audio samples using 5 bits with MLSB algorithm to reduce the number of data insertion where previously the audio sample will be compressed with Arithmetic Coding algorithm to reduce file size. In this research will also be encryption using Triple DES algorithm to better secure audio samples. The result of this research is the value of PSNR more than 50dB so it can be concluded that the image quality is still good because the value of PSNR has exceeded 40dB.
Dazert, Stefan; Thomas, Jan Peter; Büchner, Andreas; Müller, Joachim; Hempel, John Martin; Löwenheim, Hubert; Mlynski, Robert
2017-03-01
The RONDO is a single-unit cochlear implant audio processor, which omits the need for a behind-the-ear (BTE) audio processor. The primary aim was to compare speech perception results in quiet and in noise with the RONDO and the OPUS 2, a BTE audio processor. Secondary aims were to determine subjects' self-assessed levels of sound quality and gather subjective feedback on RONDO use. All speech perception tests were performed with the RONDO and the OPUS 2 behind-the-ear audio processor at 3 test intervals. Subjects were required to use the RONDO between test intervals. Subjects were tested at upgrade from the OPUS 2 to the RONDO and at 1 and 6 months after upgrade. Speech perception was determined using the Freiburg Monosyllables in quiet test and the Oldenburg Sentence Test (OLSA) in noise. Subjective perception was determined using the Hearing Implant Sound Quality Index (HISQUI 19 ), and a RONDO device-specific questionnaire. 50 subjects participated in the study. Neither speech perception scores nor self-perceived sound quality scores were significantly different at any interval between the RONDO and the OPUS 2. Subjects reported high levels of satisfaction with the RONDO. The RONDO provides comparable speech perception to the OPUS 2 while providing users with high levels of satisfaction and comfort without increasing health risk. The RONDO is a suitable and safe alternative to traditional BTE audio processors.
Computationally Efficient Clustering of Audio-Visual Meeting Data
NASA Astrophysics Data System (ADS)
Hung, Hayley; Friedland, Gerald; Yeo, Chuohao
This chapter presents novel computationally efficient algorithms to extract semantically meaningful acoustic and visual events related to each of the participants in a group discussion using the example of business meeting recordings. The recording setup involves relatively few audio-visual sensors, comprising a limited number of cameras and microphones. We first demonstrate computationally efficient algorithms that can identify who spoke and when, a problem in speech processing known as speaker diarization. We also extract visual activity features efficiently from MPEG4 video by taking advantage of the processing that was already done for video compression. Then, we present a method of associating the audio-visual data together so that the content of each participant can be managed individually. The methods presented in this article can be used as a principal component that enables many higher-level semantic analysis tasks needed in search, retrieval, and navigation.
Lin, Yu-You; Chiang, Wen-Chu; Hsieh, Ming-Ju; Sun, Jen-Tang; Chang, Yi-Chung; Ma, Matthew Huei-Ming
2018-02-01
This study aimed to conduct a systematic review and meta-analysis comparing the effect of video-assistance and audio-assistance on quality of dispatcher-instructed cardiopulmonary resuscitation (DI-CPR) for bystanders. Five databases were searched, including PubMed, Cochrane library, Embase, Scopus and NIH clinical trial, to find randomized control trials published before June 2017. Qualitative analysis and meta-analysis were undertaken to examine the difference between the quality of video-instructed and audio-instructed dispatcher-instructed bystander CPR. The database search yielded 929 records, resulting in the inclusion of 9 relevant articles in this study. Of these, 6 were included in the meta-analysis. Initiation of chest compressions was slower in the video-instructed group than in the audio-instructed group (median delay 31.5 s; 95% CI: 10.94-52.09). The difference in the number of chest compressions per minute between the groups was 19.9 (95% CI: 10.50-29.38) with significantly faster compressions in the video-instructed group than in the audio-instructed group (104.8 vs. 80.6). The odds ratio (OR) for correct hand positioning was 0.8 (95% CI: 0.53-1.30) when comparing the audio-instructed and video-instructed groups. The differences in chest compression depth (mm) and time to first ventilation (seconds) between the video-instructed group and audio-instructed group were 1.6 mm (95% CI: -8.75, 5.55) and 7.5 s (95% CI: -56.84, 71.80), respectively. Video-instructed DI-CPR significantly improved the chest compression rate compared to the audio-instructed method, and a trend for correctness of hand position was also observed. However, this method caused a delay in the commencement of bystander-initiated CPR in the simulation setting. Copyright © 2017 Elsevier B.V. All rights reserved.
NASA Astrophysics Data System (ADS)
Ndiaye, Maty; Quinquis, Catherine; Larabi, Mohamed Chaker; Le Lay, Gwenael; Saadane, Hakim; Perrine, Clency
2014-01-01
During the last decade, the important advances and widespread availability of mobile technology (operating systems, GPUs, terminal resolution and so on) have encouraged a fast development of voice and video services like video-calling. While multimedia services have largely grown on mobile devices, the generated increase of data consumption is leading to the saturation of mobile networks. In order to provide data with high bit-rates and maintain performance as close as possible to traditional networks, the 3GPP (The 3rd Generation Partnership Project) worked on a high performance standard for mobile called Long Term Evolution (LTE). In this paper, we aim at expressing recommendations related to audio and video media profiles (selection of audio and video codecs, bit-rates, frame-rates, audio and video formats) for a typical video-calling services held over LTE/4G mobile networks. These profiles are defined according to targeted devices (smartphones, tablets), so as to ensure the best possible quality of experience (QoE). Obtained results indicate that for a CIF format (352 x 288 pixels) which is usually used for smartphones, the VP8 codec provides a better image quality than the H.264 codec for low bitrates (from 128 to 384 kbps). However sequences with high motion, H.264 in slow mode is preferred. Regarding audio, better results are globally achieved using wideband codecs offering good quality except for opus codec (at 12.2 kbps).
Code of Federal Regulations, 2013 CFR
2013-10-01
... digital audio broadcasting and datacasting are authorized. The RF requirements for the DRM system are... tolerance. The frequency tolerance shall be 10 Hz. See Section 73.757(b)(2), notes 1 and 2. (3) Audio... performance of a speech codec (of the order of 3 kHz). The choice of audio quality is connected to the needs...
About subjective evaluation of adaptive video streaming
NASA Astrophysics Data System (ADS)
Tavakoli, Samira; Brunnström, Kjell; Garcia, Narciso
2015-03-01
The usage of HTTP Adaptive Streaming (HAS) technology by content providers is increasing rapidly. Having available the video content in multiple qualities, using HAS allows to adapt the quality of downloaded video to the current network conditions providing smooth video-playback. However, the time-varying video quality by itself introduces a new type of impairment. The quality adaptation can be done in different ways. In order to find the best adaptation strategy maximizing users perceptual quality it is necessary to investigate about the subjective perception of adaptation-related impairments. However, the novelties of these impairments and their comparably long time duration make most of the standardized assessment methodologies fall less suited for studying HAS degradation. Furthermore, in traditional testing methodologies, the quality of the video in audiovisual services is often evaluated separated and not in the presence of audio. Nevertheless, the requirement of jointly evaluating the audio and the video within a subjective test is a relatively under-explored research field. In this work, we address the research question of determining the appropriate assessment methodology to evaluate the sequences with time-varying quality due to the adaptation. This was done by studying the influence of different adaptation related parameters through two different subjective experiments using a methodology developed to evaluate long test sequences. In order to study the impact of audio presence on quality assessment by the test subjects, one of the experiments was done in the presence of audio stimuli. The experimental results were subsequently compared with another experiment using the standardized single stimulus Absolute Category Rating (ACR) methodology.
Alderete, John; Davies, Monica
2018-04-01
This work describes a methodology of collecting speech errors from audio recordings and investigates how some of its assumptions affect data quality and composition. Speech errors of all types (sound, lexical, syntactic, etc.) were collected by eight data collectors from audio recordings of unscripted English speech. Analysis of these errors showed that: (i) different listeners find different errors in the same audio recordings, but (ii) the frequencies of error patterns are similar across listeners; (iii) errors collected "online" using on the spot observational techniques are more likely to be affected by perceptual biases than "offline" errors collected from audio recordings; and (iv) datasets built from audio recordings can be explored and extended in a number of ways that traditional corpus studies cannot be.
High capacity reversible watermarking for audio by histogram shifting and predicted error expansion.
Wang, Fei; Xie, Zhaoxin; Chen, Zuo
2014-01-01
Being reversible, the watermarking information embedded in audio signals can be extracted while the original audio data can achieve lossless recovery. Currently, the few reversible audio watermarking algorithms are confronted with following problems: relatively low SNR (signal-to-noise) of embedded audio; a large amount of auxiliary embedded location information; and the absence of accurate capacity control capability. In this paper, we present a novel reversible audio watermarking scheme based on improved prediction error expansion and histogram shifting. First, we use differential evolution algorithm to optimize prediction coefficients and then apply prediction error expansion to output stego data. Second, in order to reduce location map bits length, we introduced histogram shifting scheme. Meanwhile, the prediction error modification threshold according to a given embedding capacity can be computed by our proposed scheme. Experiments show that this algorithm improves the SNR of embedded audio signals and embedding capacity, drastically reduces location map bits length, and enhances capacity control capability.
WebGL and web audio software lightweight components for multimedia education
NASA Astrophysics Data System (ADS)
Chang, Xin; Yuksel, Kivanc; Skarbek, Władysław
2017-08-01
The paper presents the results of our recent work on development of contemporary computing platform DC2 for multimedia education usingWebGL andWeb Audio { the W3C standards. Using literate programming paradigm the WEBSA educational tools were developed. It offers for a user (student), the access to expandable collection of WEBGL Shaders and web Audio scripts. The unique feature of DC2 is the option of literate programming, offered for both, the author and the reader in order to improve interactivity to lightweightWebGL andWeb Audio components. For instance users can define: source audio nodes including synthetic sources, destination audio nodes, and nodes for audio processing such as: sound wave shaping, spectral band filtering, convolution based modification, etc. In case of WebGL beside of classic graphics effects based on mesh and fractal definitions, the novel image processing analysis by shaders is offered like nonlinear filtering, histogram of gradients, and Bayesian classifiers.
Supervisory Control of Unmanned Vehicles
2010-04-01
than-ideal video quality (Chen et al., 2007; Chen and Thropp, 2007). Simpson et al. (2004) proposed using a spatial audio display to augment UAV...operator’s SA and discussed its utility for each of the three SA levels. They recommended that both visual and spatial audio information should be...presented concurrently. They also suggested that presenting the audio information spatially may enhance UAV operator’s sense of presence (i.e
Caffery, Liam J; Smith, Anthony C
2015-09-01
The use of fourth-generation (4G) mobile telecommunications to provide real-time video consultations were investigated in this study with the aims of determining if 4G is a suitable telecommunications technology; and secondly, to identify if variation in perceived audio and video quality were due to underlying network performance. Three patient end-points that used 4G Internet connections were evaluated. Consulting clinicians recorded their perception of audio and video quality using the International Telecommunications Union scales during clinics with these patient end-points. These scores were used to calculate a mean opinion score (MOS). The network performance metrics were obtained for each session and the relationships between these metrics and the session's quality scores were tested. Clinicians scored the quality of 50 hours of video consultations, involving 36 clinic sessions. The MOS for audio was 4.1 ± 0.62 and the MOS for video was 4.4 ± 0.22. Image impairment and effort to listen were also rated favourably. There was no correlation between audio or video quality and the network metrics of packet loss or jitter. These findings suggest that 4G networks are an appropriate telecommunication technology to deliver real-time video consultations. Variations in quality scores observed during this study were not explained by the packet loss and jitter in the underlying network. Before establishing a telemedicine service, the performance of the 4G network should be assessed at the location of the proposed service. This is due to known variability in performance of 4G networks. © The Author(s) 2015.
Exploring the Implementation of Steganography Protocols on Quantum Audio Signals
NASA Astrophysics Data System (ADS)
Chen, Kehan; Yan, Fei; Iliyasu, Abdullah M.; Zhao, Jianping
2018-02-01
Two quantum audio steganography (QAS) protocols are proposed, each of which manipulates or modifies the least significant qubit (LSQb) of the host quantum audio signal that is encoded as an FRQA (flexible representation of quantum audio) audio content. The first protocol (i.e. the conventional LSQb QAS protocol or simply the cLSQ stego protocol) is built on the exchanges between qubits encoding the quantum audio message and the LSQb of the amplitude information in the host quantum audio samples. In the second protocol, the embedding procedure to realize it implants information from a quantum audio message deep into the constraint-imposed most significant qubit (MSQb) of the host quantum audio samples, we refer to it as the pseudo MSQb QAS protocol or simply the pMSQ stego protocol. The cLSQ stego protocol is designed to guarantee high imperceptibility between the host quantum audio and its stego version, whereas the pMSQ stego protocol ensures that the resulting stego quantum audio signal is better immune to illicit tampering and copyright violations (a.k.a. robustness). Built on the circuit model of quantum computation, the circuit networks to execute the embedding and extraction algorithms of both QAS protocols are determined and simulation-based experiments are conducted to demonstrate their implementation. Outcomes attest that both protocols offer promising trade-offs in terms of imperceptibility and robustness.
Steganalysis of recorded speech
NASA Astrophysics Data System (ADS)
Johnson, Micah K.; Lyu, Siwei; Farid, Hany
2005-03-01
Digital audio provides a suitable cover for high-throughput steganography. At 16 bits per sample and sampled at a rate of 44,100 Hz, digital audio has the bit-rate to support large messages. In addition, audio is often transient and unpredictable, facilitating the hiding of messages. Using an approach similar to our universal image steganalysis, we show that hidden messages alter the underlying statistics of audio signals. Our statistical model begins by building a linear basis that captures certain statistical properties of audio signals. A low-dimensional statistical feature vector is extracted from this basis representation and used by a non-linear support vector machine for classification. We show the efficacy of this approach on LSB embedding and Hide4PGP. While no explicit assumptions about the content of the audio are made, our technique has been developed and tested on high-quality recorded speech.
The Effect of Interactive CD-ROM/Digitized Audio Courseware on Reading among Low-Literate Adults.
ERIC Educational Resources Information Center
Gretes, John A.; Green, Michael
1994-01-01
Compares a multimedia adult literacy instructional course, Reading to Educate and Develop Yourself (READY), to traditional classroom instruction by studying effects of replacing conventional learning tools with computer-assisted instruction (CD-ROMs and audio software). Results reveal that READY surpassed traditional instruction for virtually…
Studies on a Spatialized Audio Interface for Sonar
2011-10-03
addition of spatialized audio to visual displays for sonar is much akin to the development of talking movies in the early days of cinema and can be...than using the brute-force approach. PCA is one among several techniques that share similarities with the computational architecture of a
Subjective video quality evaluation of different content types under different impairments
NASA Astrophysics Data System (ADS)
Pozueco, Laura; Álvarez, Alberto; García, Xabiel; García, Roberto; Melendi, David; Díaz, Gabriel
2017-01-01
Nowadays, access to multimedia content is one of the most demanded services on the Internet. However, the transmission of audio and video over these networks is not free of problems that negatively affect user experience. Factors such as low image quality, cuts during playback or losses of audio or video, among others, can occur and there is no clear idea about the level of distortion introduced in the perceived quality. For that reason, different impairments should be evaluated based on user opinions, with the aim of analyzing the impact in the perceived quality. In this work, we carried out a subjective evaluation of different types of impairments with different types of contents, including news, cartoons, sports and action movies. A total of 100 individuals, between the ages of 20 and 68, participated in the subjective study. Results show that short-term rebuffering events negatively affect the quality of experience and that desynchronization between audio and video is the least annoying impairment. Moreover, we found that the content type determines the subjective results according to the impairment present during the playback.
Digital Audio Broadcasting in the Short Wave Bands
NASA Technical Reports Server (NTRS)
Vaisnys, Arvydas
1998-01-01
For many decades the Short Wae broadcasting service has used high power, double-sideband AM signals to reach audiences far and wide. While audio quality was usually not very high, inexpensive receivers could be used to tune into broadcasts fro distant countries.
Perception and Modeling of Affective Qualities of Musical Instrument Sounds across Pitch Registers.
McAdams, Stephen; Douglas, Chelsea; Vempala, Naresh N
2017-01-01
Composers often pick specific instruments to convey a given emotional tone in their music, partly due to their expressive possibilities, but also due to their timbres in specific registers and at given dynamic markings. Of interest to both music psychology and music informatics from a computational point of view is the relation between the acoustic properties that give rise to the timbre at a given pitch and the perceived emotional quality of the tone. Musician and nonmusician listeners were presented with 137 tones produced at a fixed dynamic marking (forte) playing tones at pitch class D# across each instrument's entire pitch range and with different playing techniques for standard orchestral instruments drawn from the brass, woodwind, string, and pitched percussion families. They rated each tone on six analogical-categorical scales in terms of emotional valence (positive/negative and pleasant/unpleasant), energy arousal (awake/tired), tension arousal (excited/calm), preference (like/dislike), and familiarity. Linear mixed models revealed interactive effects of musical training, instrument family, and pitch register, with non-linear relations between pitch register and several dependent variables. Twenty-three audio descriptors from the Timbre Toolbox were computed for each sound and analyzed in two ways: linear partial least squares regression (PLSR) and nonlinear artificial neural net modeling. These two analyses converged in terms of the importance of various spectral, temporal, and spectrotemporal audio descriptors in explaining the emotion ratings, but some differences also emerged. Different combinations of audio descriptors make major contributions to the three emotion dimensions, suggesting that they are carried by distinct acoustic properties. Valence is more positive with lower spectral slopes, a greater emergence of strong partials, and an amplitude envelope with a sharper attack and earlier decay. Higher tension arousal is carried by brighter sounds, more spectral variation and more gentle attacks. Greater energy arousal is associated with brighter sounds, with higher spectral centroids and slower decrease of the spectral slope, as well as with greater spectral emergence. The divergences between linear and nonlinear approaches are discussed.
Perception and Modeling of Affective Qualities of Musical Instrument Sounds across Pitch Registers
McAdams, Stephen; Douglas, Chelsea; Vempala, Naresh N.
2017-01-01
Composers often pick specific instruments to convey a given emotional tone in their music, partly due to their expressive possibilities, but also due to their timbres in specific registers and at given dynamic markings. Of interest to both music psychology and music informatics from a computational point of view is the relation between the acoustic properties that give rise to the timbre at a given pitch and the perceived emotional quality of the tone. Musician and nonmusician listeners were presented with 137 tones produced at a fixed dynamic marking (forte) playing tones at pitch class D# across each instrument's entire pitch range and with different playing techniques for standard orchestral instruments drawn from the brass, woodwind, string, and pitched percussion families. They rated each tone on six analogical-categorical scales in terms of emotional valence (positive/negative and pleasant/unpleasant), energy arousal (awake/tired), tension arousal (excited/calm), preference (like/dislike), and familiarity. Linear mixed models revealed interactive effects of musical training, instrument family, and pitch register, with non-linear relations between pitch register and several dependent variables. Twenty-three audio descriptors from the Timbre Toolbox were computed for each sound and analyzed in two ways: linear partial least squares regression (PLSR) and nonlinear artificial neural net modeling. These two analyses converged in terms of the importance of various spectral, temporal, and spectrotemporal audio descriptors in explaining the emotion ratings, but some differences also emerged. Different combinations of audio descriptors make major contributions to the three emotion dimensions, suggesting that they are carried by distinct acoustic properties. Valence is more positive with lower spectral slopes, a greater emergence of strong partials, and an amplitude envelope with a sharper attack and earlier decay. Higher tension arousal is carried by brighter sounds, more spectral variation and more gentle attacks. Greater energy arousal is associated with brighter sounds, with higher spectral centroids and slower decrease of the spectral slope, as well as with greater spectral emergence. The divergences between linear and nonlinear approaches are discussed. PMID:28228741
External audio for IBM-compatible computers
NASA Technical Reports Server (NTRS)
Washburn, David A.
1992-01-01
Numerous applications benefit from the presentation of computer-generated auditory stimuli at points discontiguous with the computer itself. Modification of an IBM-compatible computer for use of an external speaker is relatively easy but not intuitive. This modification is briefly described.
75 FR 25185 - Broadband Initiatives Program
Federal Register 2010, 2011, 2012, 2013, 2014
2010-05-07
..., excluding desktop or laptop computers, computer hardware and software (including anti-virus, anti-spyware, and other security software), audio or video equipment, computer network components... 10 desktop or laptop computers and individual workstations to be located within the rural library...
Design guidelines for the use of audio cues in computer interfaces
DOE Office of Scientific and Technical Information (OSTI.GOV)
Sumikawa, D.A.; Blattner, M.M.; Joy, K.I.
1985-07-01
A logical next step in the evolution of the computer-user interface is the incorporation of sound thereby using our senses of ''hearing'' in our communication with the computer. This allows our visual and auditory capacities to work in unison leading to a more effective and efficient interpretation of information received from the computer than by sight alone. In this paper we examine earcons, which are audio cues, used in the computer-user interface to provide information and feedback to the user about computer entities (these include messages and functions, as well as states and labels). The material in this paper ismore » part of a larger study that recommends guidelines for the design and use of audio cues in the computer-user interface. The complete work examines the disciplines of music, psychology, communication theory, advertising, and psychoacoustics to discover how sound is utilized and analyzed in those areas. The resulting information is organized according to the theory of semiotics, the theory of signs, into the syntax, semantics, and pragmatics of communication by sound. Here we present design guidelines for the syntax of earcons. Earcons are constructed from motives, short sequences of notes with a specific rhythm and pitch, embellished by timbre, dynamics, and register. Compound earcons and family earcons are introduced. These are related motives that serve to identify a family of related cues. Examples of earcons are given.« less
Detecting double compression of audio signal
NASA Astrophysics Data System (ADS)
Yang, Rui; Shi, Yun Q.; Huang, Jiwu
2010-01-01
MP3 is the most popular audio format nowadays in our daily life, for example music downloaded from the Internet and file saved in the digital recorder are often in MP3 format. However, low bitrate MP3s are often transcoded to high bitrate since high bitrate ones are of high commercial value. Also audio recording in digital recorder can be doctored easily by pervasive audio editing software. This paper presents two methods for the detection of double MP3 compression. The methods are essential for finding out fake-quality MP3 and audio forensics. The proposed methods use support vector machine classifiers with feature vectors formed by the distributions of the first digits of the quantized MDCT (modified discrete cosine transform) coefficients. Extensive experiments demonstrate the effectiveness of the proposed methods. To the best of our knowledge, this piece of work is the first one to detect double compression of audio signal.
Computer-based training for safety: comparing methods with older and younger workers.
Wallen, Erik S; Mulloy, Karen B
2006-01-01
Computer-based safety training is becoming more common and is being delivered to an increasingly aging workforce. Aging results in a number of changes that make it more difficult to learn from certain types of computer-based training. Instructional designs derived from cognitive learning theories may overcome some of these difficulties. Three versions of computer-based respiratory safety training were shown to older and younger workers who then took a high and a low level learning test. Younger workers did better overall. Both older and younger workers did best with the version containing text with pictures and audio narration. Computer-based training with pictures and audio narration may be beneficial for workers over 45 years of age. Computer-based safety training has advantages but workers of different ages may benefit differently. Computer-based safety programs should be designed and selected based on their ability to effectively train older as well as younger learners.
Yu, Jesang; Choi, Ji Hoon; Ma, Sun Young; Jeung, Tae Sig; Lim, Sangwook
2015-09-01
To compare audio-only biofeedback to conventional audiovisual biofeedback for regulating patients' respiration during four-dimensional radiotherapy, limiting damage to healthy surrounding tissues caused by organ movement. Six healthy volunteers were assisted by audiovisual or audio-only biofeedback systems to regulate their respirations. Volunteers breathed through a mask developed for this study by following computer-generated guiding curves displayed on a screen, combined with instructional sounds. They then performed breathing following instructional sounds only. The guiding signals and the volunteers' respiratory signals were logged at 20 samples per second. The standard deviations between the guiding and respiratory curves for the audiovisual and audio-only biofeedback systems were 21.55% and 23.19%, respectively; the average correlation coefficients were 0.9778 and 0.9756, respectively. The regularities between audiovisual and audio-only biofeedback for six volunteers' respirations were same statistically from the paired t-test. The difference between the audiovisual and audio-only biofeedback methods was not significant. Audio-only biofeedback has many advantages, as patients do not require a mask and can quickly adapt to this method in the clinic.
Cooley, Philip C.; Turner, Charles F.; O'Reilly, James M.; Allen, Danny R.; Hamill, David N.; Paddock, Richard E.
2011-01-01
This article reviews a multimedia application in the area of survey measurement research: adding audio capabilities to a computer-assisted interviewing system. Hardware and software issues are discussed, and potential hardware devices that operate from DOS platforms are reviewed. Three types of hardware devices are considered: PCMCIA devices, parallel port attachments, and laptops with built-in sound. PMID:22096271
Streaming Audio and Video: New Challenges and Opportunities for Museums.
ERIC Educational Resources Information Center
Spadaccini, Jim
Streaming audio and video present new challenges and opportunities for museums. Streaming media is easier to author and deliver to Internet audiences than ever before; digital video editing is commonplace now that the tools--computers, digital video cameras, and hard drives--are so affordable; the cost of serving video files across the Internet…
1974-09-01
introduction of modifications involving flashcards and audio have also been unsuccessful. It is felt that further progress will require a...course: Books I and 11. San Diego: Navy Personnel Research and Development Center, September 1973. Main, R. E. The effectiveness of flashcards
NASA Technical Reports Server (NTRS)
1992-01-01
Ames Research Center research into virtual reality led to the development of the Convolvotron, a high speed digital audio processing system that delivers three-dimensional sound over headphones. It consists of a two-card set designed for use with a personal computer. The Convolvotron's primary application is presentation of 3D audio signals over headphones. Four independent sound sources are filtered with large time-varying filters that compensate for motion. The perceived location of the sound remains constant. Possible applications are in air traffic control towers or airplane cockpits, hearing and perception research and virtual reality development.
Using a new, free spectrograph program to critically investigate acoustics
NASA Astrophysics Data System (ADS)
Ball, Edward; Ruiz, Michael J.
2016-11-01
We have developed an online spectrograph program with a bank of over 30 audio clips to visualise a variety of sounds. Our audio library includes everyday sounds such as speech, singing, musical instruments, birds, a baby, cat, dog, sirens, a jet, thunder, and screaming. We provide a link to a video of the sound sources superimposed with their respective spectrograms in real time. Readers can use our spectrograph program to view our library, open their own desktop audio files, and use the program in real time with a computer microphone.
Koller, Roger; Guignard, Jérémie; Caversaccio, Marco; Kompis, Martin; Senn, Pascal
2017-01-01
Background Telecommunication is limited or even impossible for more than one-thirds of all cochlear implant (CI) users. Objective We sought therefore to study the impact of voice quality on speech perception with voice over Internet protocol (VoIP) under real and adverse network conditions. Methods Telephone speech perception was assessed in 19 CI users (15-69 years, average 42 years), using the German HSM (Hochmair-Schulz-Moser) sentence test comparing Skype and conventional telephone (public switched telephone networks, PSTN) transmission using a personal computer (PC) and a digital enhanced cordless telecommunications (DECT) telephone dual device. Five different Internet transmission quality modes and four accessories (PC speakers, headphones, 3.5 mm jack audio cable, and induction loop) were compared. As a secondary outcome, the subjective perceived voice quality was assessed using the mean opinion score (MOS). Results Speech telephone perception was significantly better (median 91.6%, P<.001) with Skype compared with PSTN (median 42.5%) under optimal conditions. Skype calls under adverse network conditions (data packet loss > 15%) were not superior to conventional telephony. In addition, there were no significant differences between the tested accessories (P>.05) using a PC. Coupling a Skype DECT phone device with an audio cable to the CI, however, resulted in higher speech perception (median 65%) and subjective MOS scores (3.2) than using PSTN (median 7.5%, P<.001). Conclusions Skype calls significantly improve speech perception for CI users compared with conventional telephony under real network conditions. Listening accessories do not further improve listening experience. Current Skype DECT telephone devices do not fully offer technical advantages in voice quality. PMID:28438727
Mantokoudis, Georgios; Koller, Roger; Guignard, Jérémie; Caversaccio, Marco; Kompis, Martin; Senn, Pascal
2017-04-24
Telecommunication is limited or even impossible for more than one-thirds of all cochlear implant (CI) users. We sought therefore to study the impact of voice quality on speech perception with voice over Internet protocol (VoIP) under real and adverse network conditions. Telephone speech perception was assessed in 19 CI users (15-69 years, average 42 years), using the German HSM (Hochmair-Schulz-Moser) sentence test comparing Skype and conventional telephone (public switched telephone networks, PSTN) transmission using a personal computer (PC) and a digital enhanced cordless telecommunications (DECT) telephone dual device. Five different Internet transmission quality modes and four accessories (PC speakers, headphones, 3.5 mm jack audio cable, and induction loop) were compared. As a secondary outcome, the subjective perceived voice quality was assessed using the mean opinion score (MOS). Speech telephone perception was significantly better (median 91.6%, P<.001) with Skype compared with PSTN (median 42.5%) under optimal conditions. Skype calls under adverse network conditions (data packet loss > 15%) were not superior to conventional telephony. In addition, there were no significant differences between the tested accessories (P>.05) using a PC. Coupling a Skype DECT phone device with an audio cable to the CI, however, resulted in higher speech perception (median 65%) and subjective MOS scores (3.2) than using PSTN (median 7.5%, P<.001). Skype calls significantly improve speech perception for CI users compared with conventional telephony under real network conditions. Listening accessories do not further improve listening experience. Current Skype DECT telephone devices do not fully offer technical advantages in voice quality. ©Georgios Mantokoudis, Roger Koller, Jérémie Guignard, Marco Caversaccio, Martin Kompis, Pascal Senn. Originally published in the Journal of Medical Internet Research (http://www.jmir.org), 24.04.2017.
Improving Audio Quality in Distance Learning Applications.
ERIC Educational Resources Information Center
Richardson, Craig H.
This paper discusses common causes of problems encountered with audio systems in distance learning networks and offers practical suggestions for correcting the problems. Problems and discussions are divided into nine categories: (1) acoustics, including reverberant classrooms leading to distorted or garbled voices, as well as one-dimensional audio…
Survey data collection using Audio Computer Assisted Self-Interview.
Jones, Rachel
2003-04-01
The Audio Computer Assisted Self-Interview (ACASI) is a computer application that allows a research participant to hear survey interview items over a computer headset and read the corresponding items on a computer monitor. The ACASI automates progression from one item to the next, skipping irrelevant items. The research participant responds by pressing a number keypad, sending the data directly into a database. The ACASI was used to enhance participants' sense of privacy. A convenience sample of 257 young urban women, ages 18 to 29 years, were interviewed in neighborhood settings concerning human immune deficiency virus (HIV) sexual risk behaviors. Notebook computers were used to facilitate mobility. The overwhelming majority rated their experience with ACASI as easy to use. This article will focus on the use of ACASI in HIV behavioral research, its benefits, and approaches to resolve some identified problems with this method of data collection.
Audio fingerprint extraction for content identification
NASA Astrophysics Data System (ADS)
Shiu, Yu; Yeh, Chia-Hung; Kuo, C. C. J.
2003-11-01
In this work, we present an audio content identification system that identifies some unknown audio material by comparing its fingerprint with those extracted off-line and saved in the music database. We will describe in detail the procedure to extract audio fingerprints and demonstrate that they are robust to noise and content-preserving manipulations. The main feature in the proposed system is the zero-crossing rate extracted with the octave-band filter bank. The zero-crossing rate can be used to describe the dominant frequency in each subband with a very low computational cost. The size of audio fingerprint is small and can be efficiently stored along with the compressed files in the database. It is also robust to many modifications such as tempo change and time-alignment distortion. Besides, the octave-band filter bank is used to enhance the robustness to distortion, especially those localized on some frequency regions.
Animation, audio, and spatial ability: Optimizing multimedia for scientific explanations
NASA Astrophysics Data System (ADS)
Koroghlanian, Carol May
This study investigated the effects of audio, animation and spatial ability in a computer based instructional program for biology. The program presented instructional material via text or audio with lean text and included eight instructional sequences presented either via static illustrations or animations. High school students enrolled in a biology course were blocked by spatial ability and randomly assigned to one of four treatments (Text-Static Illustration Audio-Static Illustration, Text-Animation, Audio-Animation). The study examined the effects of instructional mode (Text vs. Audio), illustration mode (Static Illustration vs. Animation) and spatial ability (Low vs. High) on practice and posttest achievement, attitude and time. Results for practice achievement indicated that high spatial ability participants achieved more than low spatial ability participants. Similar results for posttest achievement and spatial ability were not found. Participants in the Static Illustration treatments achieved the same as participants in the Animation treatments on both the practice and posttest. Likewise, participants in the Text treatments achieved the same as participants in the Audio treatments on both the practice and posttest. In terms of attitude, participants responded favorably to the computer based instructional program. They found the program interesting, felt the static illustrations or animations made the explanations easier to understand and concentrated on learning the material. Furthermore, participants in the Animation treatments felt the information was easier to understand than participants in the Static Illustration treatments. However, no difference for any attitude item was found for participants in the Text as compared to those in the Audio treatments. Significant differences were found by Spatial Ability for three attitude items concerning concentration and interest. In all three items, the low spatial ability participants responded more positively than high spatial ability participants. In addition, low spatial ability participants reported greater mental effort than high spatial ability participants. Findings for time-in-program and time-in-instruction indicated that participants in the Animation treatments took significantly more time than participants in the Static Illustration treatments. No time differences of any type were found for participants in the Text versus Audio treatments. Implications for the design of multimedia instruction and topics for future research are included in the discussion.
Student Preferences for Online Lecture Formats: Does Prior Experience Matter?
ERIC Educational Resources Information Center
Drouin, Michelle; Hile, Rachel E.; Vartanian, Lesa R.; Webb, Janae
2013-01-01
We examined undergraduate students' quality ratings of and preferences for different types of online lecture formats. Students preferred richer online lecture formats that included both audio and visual components; however, there were no significant differences between students' ratings of PowerPoint lectures with "audio" of the…
Audio/ Videoconferencing Packages: Low Cost
ERIC Educational Resources Information Center
Treblay, Remy; Fyvie, Barb; Koritko, Brenda
2005-01-01
A comparison was conducted of "Voxwire MeetingRoom" and "iVocalize" v4.1.0.3, both Web-conferencing products using voice-over-Internet protocol (VoIP) to provide unlimited, inexpensive, international audio communication, and high-quality Web-conferencing fostering collaborative learning. The study used the evaluation criteria used in earlier…
Radioactive Decay: Audio Data Collection
ERIC Educational Resources Information Center
Struthers, Allan
2009-01-01
Many phenomena generate interesting audible time series. This data can be collected and processed using audio software. The free software package "Audacity" is used to demonstrate the process by recording, processing, and extracting click times from an inexpensive radiation detector. The high quality of the data is demonstrated with a simple…
ERIC Educational Resources Information Center
Li, Chenxi; Wu, Ligao; Li, Chen; Tang, Jinlan
2017-01-01
This work-in-progress doctoral research project aims to identify meaning negotiation patterns in synchronous audio and video Computer-Mediated Communication (CMC) environments based on the model of CMC text chat proposed by Smith (2003). The study was conducted in the Institute of Online Education at Beijing Foreign Studies University. Four dyads…
Marschall-Lévesque, Shawn; Rouleau, Joanne-Lucine; Renaud, Patrice
2018-02-01
Penile plethysmography (PPG) is a measure of sexual interests that relies heavily on the stimuli it uses to generate valid results. Ethical considerations surrounding the use of real images in PPG have further limited the content admissible for these stimuli. To palliate this limitation, the current study aimed to combine audio and visual stimuli by incorporating computer-generated characters to create new stimuli capable of accurately classifying sex offenders with child victims, while also increasing the number of valid profiles. Three modalities (audio, visual, and audiovisual) were compared using two groups (15 sex offenders with child victims and 15 non-offenders). Both the new visual and audiovisual stimuli resulted in a 13% increase in the number of valid profiles at 2.5 mm, when compared to the standard audio stimuli. Furthermore, the new audiovisual stimuli generated a 34% increase in penile responses. All three modalities were able to discriminate between the two groups by their responses to the adult and child stimuli. Lastly, sexual interest indices for all three modalities could accurately classify participants in their appropriate groups, as demonstrated by ROC curve analysis (i.e., audio AUC = .81, 95% CI [.60, 1.00]; visual AUC = .84, 95% CI [.66, 1.00], and audiovisual AUC = .83, 95% CI [.63, 1.00]). Results suggest that computer-generated characters allow accurate discrimination of sex offenders with child victims and can be added to already validated stimuli to increase the number of valid profiles. The implications of audiovisual stimuli using computer-generated characters and their possible use in PPG evaluations are also discussed.
Wang, Nancy X. R.; Olson, Jared D.; Ojemann, Jeffrey G.; Rao, Rajesh P. N.; Brunton, Bingni W.
2016-01-01
Fully automated decoding of human activities and intentions from direct neural recordings is a tantalizing challenge in brain-computer interfacing. Implementing Brain Computer Interfaces (BCIs) outside carefully controlled experiments in laboratory settings requires adaptive and scalable strategies with minimal supervision. Here we describe an unsupervised approach to decoding neural states from naturalistic human brain recordings. We analyzed continuous, long-term electrocorticography (ECoG) data recorded over many days from the brain of subjects in a hospital room, with simultaneous audio and video recordings. We discovered coherent clusters in high-dimensional ECoG recordings using hierarchical clustering and automatically annotated them using speech and movement labels extracted from audio and video. To our knowledge, this represents the first time techniques from computer vision and speech processing have been used for natural ECoG decoding. Interpretable behaviors were decoded from ECoG data, including moving, speaking and resting; the results were assessed by comparison with manual annotation. Discovered clusters were projected back onto the brain revealing features consistent with known functional areas, opening the door to automated functional brain mapping in natural settings. PMID:27148018
Yu, Jesang; Choi, Ji Hoon; Ma, Sun Young; Jeung, Tae Sig
2015-01-01
Purpose To compare audio-only biofeedback to conventional audiovisual biofeedback for regulating patients' respiration during four-dimensional radiotherapy, limiting damage to healthy surrounding tissues caused by organ movement. Materials and Methods Six healthy volunteers were assisted by audiovisual or audio-only biofeedback systems to regulate their respirations. Volunteers breathed through a mask developed for this study by following computer-generated guiding curves displayed on a screen, combined with instructional sounds. They then performed breathing following instructional sounds only. The guiding signals and the volunteers' respiratory signals were logged at 20 samples per second. Results The standard deviations between the guiding and respiratory curves for the audiovisual and audio-only biofeedback systems were 21.55% and 23.19%, respectively; the average correlation coefficients were 0.9778 and 0.9756, respectively. The regularities between audiovisual and audio-only biofeedback for six volunteers' respirations were same statistically from the paired t-test. Conclusion The difference between the audiovisual and audio-only biofeedback methods was not significant. Audio-only biofeedback has many advantages, as patients do not require a mask and can quickly adapt to this method in the clinic. PMID:26484309
Direct broadcast satellite-radio market, legal, regulatory, and business considerations
NASA Technical Reports Server (NTRS)
Sood, Des R.
1991-01-01
A Direct Broadcast Satellite-Radio (DBS-R) System offers the prospect of delivering high quality audio broadcasts to large audiences at costs lower than or comparable to those incurred using the current means of broadcasting. The maturation of mobile communications technologies, and advances in microelectronics and digital signal processing now make it possible to bring this technology to the marketplace. Heightened consumer interest in improved audio quality coupled with the technological and economic feasibility of meeting this demand via DBS-R make it opportune to start planning for implementation of DBS-R Systems. NASA-Lewis and the Voice of America as part of their on-going efforts to improve the quality of international audio broadcasts, have undertaken a number of tasks to more clearly define the technical, marketing, organizational, legal, and regulatory issues underlying implementation of DBS-R Systems. The results and an assessment is presented of the business considerations underlying the construction, launch, and operation of DBS-R Systems.
Direct broadcast satellite-radio market, legal, regulatory, and business considerations
NASA Astrophysics Data System (ADS)
Sood, Des R.
1991-03-01
A Direct Broadcast Satellite-Radio (DBS-R) System offers the prospect of delivering high quality audio broadcasts to large audiences at costs lower than or comparable to those incurred using the current means of broadcasting. The maturation of mobile communications technologies, and advances in microelectronics and digital signal processing now make it possible to bring this technology to the marketplace. Heightened consumer interest in improved audio quality coupled with the technological and economic feasibility of meeting this demand via DBS-R make it opportune to start planning for implementation of DBS-R Systems. NASA-Lewis and the Voice of America as part of their on-going efforts to improve the quality of international audio broadcasts, have undertaken a number of tasks to more clearly define the technical, marketing, organizational, legal, and regulatory issues underlying implementation of DBS-R Systems. The results and an assessment is presented of the business considerations underlying the construction, launch, and operation of DBS-R Systems.
Experienced quality factors: qualitative evaluation approach to audiovisual quality
NASA Astrophysics Data System (ADS)
Jumisko-Pyykkö, Satu; Häkkinen, Jukka; Nyman, Göte
2007-02-01
Subjective evaluation is used to identify impairment factors of multimedia quality. The final quality is often formulated via quantitative experiments, but this approach has its constraints, as subject's quality interpretations, experiences and quality evaluation criteria are disregarded. To identify these quality evaluation factors, this study examined qualitatively the criteria participants used to evaluate audiovisual video quality. A semi-structured interview was conducted with 60 participants after a subjective audiovisual quality evaluation experiment. The assessment compared several, relatively low audio-video bitrate ratios with five different television contents on mobile device. In the analysis, methodological triangulation (grounded theory, Bayesian networks and correspondence analysis) was applied to approach the qualitative quality. The results showed that the most important evaluation criteria were the factors of visual quality, contents, factors of audio quality, usefulness - followability and audiovisual interaction. Several relations between the quality factors and the similarities between the contents were identified. As a research methodological recommendation, the focus on content and usage related factors need to be further examined to improve the quality evaluation experiments.
Quality models for audiovisual streaming
NASA Astrophysics Data System (ADS)
Thang, Truong Cong; Kim, Young Suk; Kim, Cheon Seog; Ro, Yong Man
2006-01-01
Quality is an essential factor in multimedia communication, especially in compression and adaptation. Quality metrics can be divided into three categories: within-modality quality, cross-modality quality, and multi-modality quality. Most research has so far focused on within-modality quality. Moreover, quality is normally just considered from the perceptual perspective. In practice, content may be drastically adapted, even converted to another modality. In this case, we should consider the quality from semantic perspective as well. In this work, we investigate the multi-modality quality from the semantic perspective. To model the semantic quality, we apply the concept of "conceptual graph", which consists of semantic nodes and relations between the nodes. As an typical of multi-modality example, we focus on audiovisual streaming service. Specifically, we evaluate the amount of information conveyed by a audiovisual content where both video and audio channels may be strongly degraded, even audio are converted to text. In the experiments, we also consider the perceptual quality model of audiovisual content, so as to see the difference with semantic quality model.
Automatic violence detection in digital movies
NASA Astrophysics Data System (ADS)
Fischer, Stephan
1996-11-01
Research on computer-based recognition of violence is scant. We are working on the automatic recognition of violence in digital movies, a first step towards the goal of a computer- assisted system capable of protecting children against TV programs containing a great deal of violence. In the video domain a collision detection and a model-mapping to locate human figures are run, while the creation and comparison of fingerprints to find certain events are run int he audio domain. This article centers on the recognition of fist- fights in the video domain and on the recognition of shots, explosions and cries in the audio domain.
Audio Feedback -- Better Feedback?
ERIC Educational Resources Information Center
Voelkel, Susanne; Mello, Luciane V.
2014-01-01
National Student Survey (NSS) results show that many students are dissatisfied with the amount and quality of feedback they get for their work. This study reports on two case studies in which we tried to address these issues by introducing audio feedback to one undergraduate (UG) and one postgraduate (PG) class, respectively. In case study one…
Culturally Diverse Videos, Audios, and CD-ROMs for Children and Young Adults.
ERIC Educational Resources Information Center
Wood, Irene
The purpose of this book is to help librarians develop high quality video, audio, and CD-ROM collections for preschool through high school learning with titles that reflect the ethnic heritage and experience of the diverse North American population, primarily African Americans, Asian Americans, Hispanic Americans, and Native Americans. The more…
The Changing Role of the Educational Video in Higher Distance Education
ERIC Educational Resources Information Center
Laaser, Wolfram; Toloza, Eduardo A.
2017-01-01
The article argues that the ongoing usage of audio visual media is falling behind in terms of educational quality compared to prior achievements in the history of distance education. After reviewing some important steps and experiences of audio visual digital media development, we analyse predominant presentation formats on the Web. Special focus…
NASA Astrophysics Data System (ADS)
Zhang, Y.; Paulson, K. V.
For audio-frequency magnetotelluric surveys where the signals are lightning-stroke transients, the conventional Fourier transform method often fails to produce a high quality impedance tensor. An alternative approach is to use the wavelet transform method which is capable of localizing target information simultaneously in both the temporal and frequency domains. Unlike Fourier analysis that yields an average amplitude and phase, the wavelet transform produces an instantaneous estimate of the amplitude and phase of a signal. In this paper a complex well-localized wavelet, the Morlet wavelet, has been used to transform and analyze audio-frequency magnetotelluric data. With the Morlet wavelet, the magnetotelluric impedance tensor can be computed directly in the wavelet transform domain. The lightning-stroke transients are easily identified on the dilation-translation plane. Choosing those wavelet transform values where the signals are located, a higher signal-to-noise ratio estimation of the impedance tensor can be obtained. In a test using real data, the wavelet transform showed a significant improvement in the signal-to-noise ratio over the conventional Fourier transform.
Spatial domain entertainment audio decompression/compression
NASA Astrophysics Data System (ADS)
Chan, Y. K.; Tam, Ka Him K.
2014-02-01
The ARM7 NEON processor with 128bit SIMD hardware accelerator requires a peak performance of 13.99 Mega Cycles per Second for MP3 stereo entertainment quality decoding. For similar compression bit rate, OGG and AAC is preferred over MP3. The Patent Cooperation Treaty Application dated 28/August/2012 describes an audio decompression scheme producing a sequence of interleaving "min to Max" and "Max to min" rising and falling segments. The number of interior audio samples bound by "min to Max" or "Max to min" can be {0|1|…|N} audio samples. The magnitudes of samples, including the bounding min and Max, are distributed as normalized constants within the 0 and 1 of the bounding magnitudes. The decompressed audio is then a "sequence of static segments" on a frame by frame basis. Some of these frames needed to be post processed to elevate high frequency. The post processing is compression efficiency neutral and the additional decoding complexity is only a small fraction of the overall decoding complexity without the need of extra hardware. Compression efficiency can be speculated as very high as source audio had been decimated and converted to a set of data with only "segment length and corresponding segment magnitude" attributes. The PCT describes how these two attributes are efficiently coded by the PCT innovative coding scheme. The PCT decoding efficiency is obviously very high and decoding latency is basically zero. Both hardware requirement and run time is at least an order of magnitude better than MP3 variants. The side benefit is ultra low power consumption on mobile device. The acid test on how such a simplistic waveform representation can indeed reproduce authentic decompressed quality is benchmarked versus OGG(aoTuv Beta 6.03) by three pair of stereo audio frames and one broadcast like voice audio frame with each frame consisting 2,028 samples at 44,100KHz sampling frequency.
For Kids, by Kids: Our City Podcast
ERIC Educational Resources Information Center
Vincent, Tony; van't Hooft, Mark
2007-01-01
In this article, the authors discuss podcasting and provide ways on how to create podcasts. A podcast is an audio or video file that is posted on the web that can easily be cataloged and automatically downloaded to a computer or mobile device capable of playing back audio or video files. Podcasting is a powerful tool for educators to get students…
Podcasting: A Preliminary Classroom Study
ERIC Educational Resources Information Center
Aristizabal, Alexander
2009-01-01
Podcasting is a term introduced through the use of Apple Computer, Inc.'s iPod, a term which denotes how a portable audio player can be used to download audio files, mostly MP3s, and be heard at the user's convenience. Initially such an operation was intended for entertainment; however, it has proven itself to be an important tool in the field of…
ERIC Educational Resources Information Center
Bell, Adam Patrick
2015-01-01
The proliferation of computers, tablets, and smartphones has resulted in digital audio workstations (DAWs) such as GarageBand in being some of the most widely distributed musical instruments. Positing that software designers are dictating the music education of DAW-dependent music-makers, I examine the fallacy that music-making applications such…
Highlight summarization in golf videos using audio signals
NASA Astrophysics Data System (ADS)
Kim, Hyoung-Gook; Kim, Jin Young
2008-01-01
In this paper, we present an automatic summarization of highlights in golf videos based on audio information alone without video information. The proposed highlight summarization system is carried out based on semantic audio segmentation and detection on action units from audio signals. Studio speech, field speech, music, and applause are segmented by means of sound classification. Swing is detected by the methods of impulse onset detection. Sounds like swing and applause form a complete action unit, while studio speech and music parts are used to anchor the program structure. With the advantage of highly precise detection of applause, highlights are extracted effectively. Our experimental results obtain high classification precision on 18 golf games. It proves that the proposed system is very effective and computationally efficient to apply the technology to embedded consumer electronic devices.
Case Study: Audio-Guided Learning, with Computer Graphics.
ERIC Educational Resources Information Center
Koumi, Jack; Daniels, Judith
1994-01-01
Describes teaching packages which involve the use of audiotape recordings with personal computers in Open University (United Kingdom) mathematics courses. Topics addressed include software development; computer graphics; pedagogic principles for distance education; feedback, including course evaluations and student surveys; and future plans.…
Audio-visual temporal perception in children with restored hearing.
Gori, Monica; Chilosi, Anna; Forli, Francesca; Burr, David
2017-05-01
It is not clear how audio-visual temporal perception develops in children with restored hearing. In this study we measured temporal discrimination thresholds with an audio-visual temporal bisection task in 9 deaf children with restored audition, and 22 typically hearing children. In typically hearing children, audition was more precise than vision, with no gain in multisensory conditions (as previously reported in Gori et al. (2012b)). However, deaf children with restored audition showed similar thresholds for audio and visual thresholds and some evidence of gain in audio-visual temporal multisensory conditions. Interestingly, we found a strong correlation between auditory weighting of multisensory signals and quality of language: patients who gave more weight to audition had better language skills. Similarly, auditory thresholds for the temporal bisection task were also a good predictor of language skills. This result supports the idea that the temporal auditory processing is associated with language development. Copyright © 2017. Published by Elsevier Ltd.
Worldwide survey of direct-to-listener digital audio delivery systems development since WARC-1992
NASA Technical Reports Server (NTRS)
Messer, Dion D.
1993-01-01
Each country was allocated frequency band(s) for direct-to-listener digital audio broadcasting at WARC-92. These allocations were near 1500, 2300, and 2600 MHz. In addition, some countries are encouraging the development of digital audio broadcasting services for terrestrial delivery only in the VHF bands (at frequencies from roughly 50 to 300 MHz) and in the medium-wave broadcasting band (AM band) (from roughly 0.5 to 1.7 MHz). The development activity increase was explosive. Current development, as of February 1993, as it is known to the author is summarized. The information given includes the following characteristics, as appropriate, for each planned system: coverage areas, audio quality, number of audio channels, delivery via satellite/terrestrial or both, carrier frequency bands, modulation methods, source coding, and channel coding. Most proponents claim that they will be operational in 3 or 4 years.
Effects of aging on audio-visual speech integration.
Huyse, Aurélie; Leybaert, Jacqueline; Berthommier, Frédéric
2014-10-01
This study investigated the impact of aging on audio-visual speech integration. A syllable identification task was presented in auditory-only, visual-only, and audio-visual congruent and incongruent conditions. Visual cues were either degraded or unmodified. Stimuli were embedded in stationary noise alternating with modulated noise. Fifteen young adults and 15 older adults participated in this study. Results showed that older adults had preserved lipreading abilities when the visual input was clear but not when it was degraded. The impact of aging on audio-visual integration also depended on the quality of the visual cues. In the visual clear condition, the audio-visual gain was similar in both groups and analyses in the framework of the fuzzy-logical model of perception confirmed that older adults did not differ from younger adults in their audio-visual integration abilities. In the visual reduction condition, the audio-visual gain was reduced in the older group, but only when the noise was stationary, suggesting that older participants could compensate for the loss of lipreading abilities by using the auditory information available in the valleys of the noise. The fuzzy-logical model of perception confirmed the significant impact of aging on audio-visual integration by showing an increased weight of audition in the older group.
NASA Astrophysics Data System (ADS)
Cerwin, Steve; Barnes, Julie; Kell, Scott; Walters, Mark
2003-09-01
This paper describes development and application of a novel method to accomplish real-time solid angle acoustic direction finding using two 8-element orthogonal microphone arrays. The developed prototype system was intended for localization and signature recognition of ground-based sounds from a small UAV. Recent advances in computer speeds have enabled the implementation of microphone arrays in many audio applications. Still, the real-time presentation of a two-dimensional sound field for the purpose of audio target localization is computationally challenging. In order to overcome this challenge, a crosspower spectrum phase1 (CSP) technique was applied to each 8-element arm of a 16-element cross array to provide audio target localization. In this paper, we describe the technique and compare it with two other commonly used techniques; Cross-Spectral Matrix2 and MUSIC3. The results show that the CSP technique applied to two 8-element orthogonal arrays provides a computationally efficient solution with reasonable accuracy and tolerable artifacts, sufficient for real-time applications. Additional topics include development of a synchronized 16-channel transmitter and receiver to relay the airborne data to the ground-based processor and presentation of test data demonstrating both ground-mounted operation and airborne localization of ground-based gunshots and loud engine sounds.
Raffaelli, Marcela; Armstrong, Jessica; Tran, Steve P; Griffith, Aisha N; Walker, Kathrin; Gutierrez, Vanessa
2016-06-01
Computer-assisted data collection offers advantages over traditional paper and pencil measures; however, little guidance is available regarding the logistics of conducting computer-assisted data collection with adolescents in group settings. To address this gap, we draw on our experiences conducting a multi-site longitudinal study of adolescent development. Structured questionnaires programmed on laptop computers using Audio Computer Assisted Self-Interviewing (ACASI) were administered to groups of adolescents in community-based and afterschool programs. Although implementing ACASI required additional work before entering the field, we benefited from reduced data processing time, high data quality, and high levels of youth motivation. Preliminary findings from an ethnically diverse sample of 265 youth indicate favorable perceptions of using ACASI. Using our experiences as a case study, we provide recommendations on selecting an appropriate data collection device (including hardware and software), preparing and testing the ACASI, conducting data collection in the field, and managing data. Copyright © 2016 The Foundation for Professionals in Services for Adolescents. Published by Elsevier Ltd. All rights reserved.
Podcasting: contemporary patient education.
Abreu, Daniel V; Tamura, Thomas K; Sipp, J Andrew; Keamy, Donald G; Eavey, Roland D
2008-04-01
Portable video technology is a widely available new tool with potential to be used by pediatric otolaryngology practices for patient and family education. Podcasts are media broadcasts that employ this new technology. They can be accessed via the Internet and viewed either on a personal computer or on a handheld device, such as an iPod or an MP3 player. We wished to examine the feasibility of establishing a podcast-hosting Web site. We digitally recorded pediatric otologic procedures in the operating room and saved the digital files to DVDs. We then edited the DVDs at home with video-editing software on a personal computer. Next, spoken narrative was recorded with audio-recording software and combined with the edited video clips. The final products were converted into the M4V file format, and the final versions were uploaded onto our hospital's Web site. We then downloaded the podcasts onto a high-quality portable media player so that we could evaluate their quality. All of the podcasts are now on the hospital Web site, where they can be downloaded by patients and families at no cost. The site includes instructions on how to download the appropriate free software for viewing the podcasts on a portable media player or on a computer. Using this technology for patient education expands the audience and permits portability of information. We conclude that a home computer can be used to inexpensively create informative surgery demonstrations that can be accessed via a Web site and transferred to portable viewing devices with excellent quality.
University of Arizona: College and University Systems Environment.
ERIC Educational Resources Information Center
CAUSE/EFFECT, 1985
1985-01-01
The University of Arizona has begun to reorganize campus computing. Six working groups were formed to address six areas of computing: academic computing, library automation, administrative data processing and information systems, writing and graphics, video and audio services, and outreach and public service. (MLW)
Computer Networking with the Victorian Correspondence School.
ERIC Educational Resources Information Center
Conboy, Ian
During 1985 the Education Department installed two-way radios in 44 remote secondary schools in Victoria, Australia, to improve turn-around time for correspondence assignments. Subsequently, teacher supervisors at Melbourne's Correspondence School sought ways to further augument audio interactivity with computer networking. Computer equipment was…
Designing sound and visual components for enhancement of urban soundscapes.
Hong, Joo Young; Jeon, Jin Yong
2013-09-01
The aim of this study is to investigate the effect of audio-visual components on environmental quality to improve soundscape. Natural sounds with road traffic noise and visual components in urban streets were evaluated through laboratory experiments. Waterfall and stream water sounds, as well as bird sounds, were selected to enhance the soundscape. Sixteen photomontages of a streetscape were constructed in combination with two types of water features and three types of vegetation which were chosen as positive visual components. The experiments consisted of audio-only, visual-only, and audio-visual conditions. The preferences and environmental qualities of the stimuli were evaluated by a numerical scale and 12 pairs of adjectives, respectively. The results showed that bird sounds were the most preferred among the natural sounds, while the sound of falling water was found to degrade the soundscape quality when the road traffic noise level was high. The visual effects of vegetation on aesthetic preference were significant, but those of water features relatively small. It was revealed that the perceptual dimensions of the environment were different from the noise levels. Particularly, the acoustic comfort factor related to soundscape quality considerably influenced preference for the overall environment at a higher level of road traffic noise.
ERIC Educational Resources Information Center
Robinson, David E.
1997-01-01
One solution to poor quality sound in student video projects is a four-track audio cassette recorder. This article discusses the advantages of four-track over single-track recorders and compares two student productions, one using a single-track and the other a four-track recorder. (PEN)
ERIC Educational Resources Information Center
Moon, Donald K.
This document is one in a series of reports which reviews instructional materials and equipment and offers suggestions about how to select equipment. Topics discussed include: (1) the general criteria for audio-visual equipment selection such as performance, safety, comparability, sturdiness and repairability; and (2) specific equipment criteria…
ERIC Educational Resources Information Center
Macmullen, Paul
The main focus of this document is on audioconferencing, which in distance education contexts provides "virtual" interaction equivalent in quality to face-to-face, conventional classroom interaction. The applications of audiotape and audio broadcast are covered only briefly. Discussion first includes reasons for using audioconferencing…
Constructing a Streaming Video-Based Learning Forum for Collaborative Learning
ERIC Educational Resources Information Center
Chang, Chih-Kai
2004-01-01
As web-based courses using videos have become popular in recent years, the issue of managing audio-visual aids has become pertinent. Generally, the contents of audio-visual aids may include a lecture, an interview, a report, or an experiment, which may be transformed into a streaming format capable of making the quality of Internet-based videos…
Digital Documentation: Using Computers to Create Multimedia Reports.
ERIC Educational Resources Information Center
Speitel, Tom; And Others
1996-01-01
Describes methods for creating integrated multimedia documents using recent advances in print, audio, and video digitization that bring added usefulness to computers as data acquisition, processing, and presentation tools. Discusses advantages of digital documentation. (JRH)
Federal Register 2010, 2011, 2012, 2013, 2014
2012-05-23
... methods of administration (e.g., computer assisted personal interviews [CAPI], audio computer assisted self-interviews [ACASI], web-based interviews). Cognitive testing of these materials and methods will...
Advances in audio source seperation and multisource audio content retrieval
NASA Astrophysics Data System (ADS)
Vincent, Emmanuel
2012-06-01
Audio source separation aims to extract the signals of individual sound sources from a given recording. In this paper, we review three recent advances which improve the robustness of source separation in real-world challenging scenarios and enable its use for multisource content retrieval tasks, such as automatic speech recognition (ASR) or acoustic event detection (AED) in noisy environments. We present a Flexible Audio Source Separation Toolkit (FASST) and discuss its advantages compared to earlier approaches such as independent component analysis (ICA) and sparse component analysis (SCA). We explain how cues as diverse as harmonicity, spectral envelope, temporal fine structure or spatial location can be jointly exploited by this toolkit. We subsequently present the uncertainty decoding (UD) framework for the integration of audio source separation and audio content retrieval. We show how the uncertainty about the separated source signals can be accurately estimated and propagated to the features. Finally, we explain how this uncertainty can be efficiently exploited by a classifier, both at the training and the decoding stage. We illustrate the resulting performance improvements in terms of speech separation quality and speaker recognition accuracy.
Nonlinear dynamic macromodeling techniques for audio systems
NASA Astrophysics Data System (ADS)
Ogrodzki, Jan; Bieńkowski, Piotr
2015-09-01
This paper develops a modelling method and a models identification technique for the nonlinear dynamic audio systems. Identification is performed by means of a behavioral approach based on a polynomial approximation. This approach makes use of Discrete Fourier Transform and Harmonic Balance Method. A model of an audio system is first created and identified and then it is simulated in real time using an algorithm of low computational complexity. The algorithm consists in real time emulation of the system response rather than in simulation of the system itself. The proposed software is written in Python language using object oriented programming techniques. The code is optimized for a multithreads environment.
Implementation of Audio Computer-Assisted Interviewing Software in HIV/AIDS Research
Pluhar, Erika; Yeager, Katherine A.; Corkran, Carol; McCarty, Frances; Holstad, Marcia McDonnell; Denzmore-Nwagbara, Pamela; Fielder, Bridget; DiIorio, Colleen
2007-01-01
Computer assisted interviewing (CAI) has begun to play a more prominent role in HIV/AIDS prevention research. Despite the increased popularity of CAI, particularly audio computer assisted self-interviewing (ACASI), some research teams are still reluctant to implement ACASI technology due to lack of familiarity with the practical issues related to using these software packages. The purpose of this paper is to describe the implementation of one particular ACASI software package, the Questionnaire Development System™ (QDS™), in several nursing and HIV/AIDS prevention research settings. We present acceptability and satisfaction data from two large-scale public health studies in which we have used QDS with diverse populations. We also address issues related to developing and programming a questionnaire, discuss practical strategies related to planning for and implementing ACASI in the field, including selecting equipment, training staff, and collecting and transferring data, and summarize advantages and disadvantages of computer assisted research methods. PMID:17662924
Handels, H; Busch, C; Encarnação, J; Hahn, C; Kühn, V; Miehe, J; Pöppl, S I; Rinast, E; Rossmanith, C; Seibert, F; Will, A
1997-03-01
The software system KAMEDIN (Kooperatives Arbeiten und MEdizinische Diagnostik auf Innovativen Netzen) is a multimedia telemedicine system for exchange, cooperative diagnostics, and remote analysis of digital medical image data. It provides components for visualisation, processing, and synchronised audio-visual discussion of medical images. Techniques of computer supported cooperative work (CSCW) synchronise user interactions during a teleconference. Visibility of both local and remote cursor on the conference workstations facilitates telepointing and reinforces the conference partner's telepresence. Audio communication during teleconferences is supported by an integrated audio component. Furthermore, brain tissue segmentation with artificial neural networks can be performed on an external supercomputer as a remote image analysis procedure. KAMEDIN is designed as a low cost CSCW tool for ISDN based telecommunication. However it can be used on any TCP/IP supporting network. In a field test, KAMEDIN was installed in 15 clinics and medical departments to validate the systems' usability. The telemedicine system KAMEDIN has been developed, tested, and evaluated within a research project sponsored by German Telekom.
Speaker Localisation Using Time Difference of Arrival
2008-04-01
School of Electrical and Electronic Engineering of the University of Adelaide. His area of expertise and interest is in Signal Processing including audio ...support of Theatre intelligence capabilities. His recent research interests include: information visualisation , text and data mining, and speech and...by: steering microphone arrays to improve the quality of audio pickup for recording, communication and transcription; enhancing the separation – and
Developing a Framework for Effective Audio Feedback: A Case Study
ERIC Educational Resources Information Center
Hennessy, Claire; Forrester, Gillian
2014-01-01
The increase in the use of technology-enhanced learning in higher education has included a growing interest in new approaches to enhance the quality of feedback given to students. Audio feedback is one method that has become more popular, yet evaluating its role in feedback delivery is still an emerging area for research. This paper is based on a…
"Are You Listening Please?" The Advantages of Electronic Audio Feedback Compared to Written Feedback
ERIC Educational Resources Information Center
Lunt, Tom; Curran, John
2010-01-01
Feedback on students' work is, probably, one of the most important aspects of learning, yet students' report, according to the National Union of Students (NUS) Survey of 2008, unhappiness with the feedback process. Students were unhappy with the quality, detail and timing of feedback. This paper examines the benefits of using audio, as opposed to…
Integrated Spacesuit Audio System Enhances Speech Quality and Reduces Noise
NASA Technical Reports Server (NTRS)
Huang, Yiteng Arden; Chen, Jingdong; Chen, Shaoyan Sharyl
2009-01-01
A new approach has been proposed for increasing astronaut comfort and speech capture. Currently, the special design of a spacesuit forms an extreme acoustic environment making it difficult to capture clear speech without compromising comfort. The proposed Integrated Spacesuit Audio (ISA) system is to incorporate the microphones into the helmet and use software to extract voice signals from background noise.
ERIC Educational Resources Information Center
Lockwood, Nicholas S.
2011-01-01
Geographically dispersed teams rely on information and communication technologies (ICTs) to communicate and collaborate. Three ICTs that have received attention are audio conferencing (AC), video conferencing (VC), and, recently, 3D virtual environments (3D VEs). These ICTs offer modes of communication that differ primarily in the number and type…
2006-03-31
from existing image steganography and steganalysis techniques, the overall objective of Task (b) is to design and implement audio steganography in...general design of the VoIP steganography algorithm is based on known LSB hiding techniques (used for example in StegHide (http...system. Nasir Memon et. al. described a steganalyzer based on image quality metrics [AMS03]. Basically, the main idea to detect steganography by
Capacity-optimized mp2 audio watermarking
NASA Astrophysics Data System (ADS)
Steinebach, Martin; Dittmann, Jana
2003-06-01
Today a number of audio watermarking algorithms have been proposed, some of them at a quality making them suitable for commercial applications. The focus of most of these algorithms is copyright protection. Therefore, transparency and robustness are the most discussed and optimised parameters. But other applications for audio watermarking can also be identified stressing other parameters like complexity or payload. In our paper, we introduce a new mp2 audio watermarking algorithm optimised for high payload. Our algorithm uses the scale factors of an mp2 file for watermark embedding. They are grouped and masked based on a pseudo-random pattern generated from a secret key. In each group, we embed one bit. Depending on the bit to embed, we change the scale factors by adding 1 where necessary until it includes either more even or uneven scale factors. An uneven group has a 1 embedded, an even group a 0. The same rule is later applied to detect the watermark. The group size can be increased or decreased for transparency/payload trade-off. We embed 160 bits or more in an mp2 file per second without reducing perceived quality. As an application example, we introduce a prototypic Karaoke system displaying song lyrics embedded as a watermark.
ERIC Educational Resources Information Center
Aarnio, Pauliina; Kulmala, Teija
2016-01-01
Self-interview methods such as audio computer-assisted self-interviewing (ACASI) are used to improve the accuracy of interview data on sensitive topics in large trials. Small field studies on sensitive topics would benefit from methodological alternatives. In a study on male involvement in antenatal HIV testing in a largely illiterate population…
Report on Distance Learning Technologies.
1995-09-01
26 cities. The CSX system includes full-motion video, animations , audio, and interactive examples and testing to teach the use of a new computer...video. The change to all-digital media now permits the use of full-motion video, animation , and audio on networks. It is possible to have independent...is possible to download entire multimedia presentations from the network. To date there is not a great deal known about teaching courses using the
Reproductive and Hormonal Risk Factors for Breast Cancer in Blind Women
2008-08-01
informational letters sent directly from the Perkins Braille and Talking Book Library in Watertown Massachusetts, the ACB and the Canadian National Institute...choice. Survey formats included via e-mail, website, compact disc, computer disk, audio tape, large print, Braille , in person, or verbally over the...survey data during telephone interviews or following receipt of written or audio surveys. Braille surveys were transcribed by a third party and
Quo Vadimus? The 21st Century and Multimedia.
ERIC Educational Resources Information Center
Kuhn, Allan D.
This paper relates the concept of computer-driven multimedia to the National Aeronautics and Space Administration (NASA) Scientific and Technical Information Program (STIP). Multimedia is defined here as computer integration and output of text, animation, audio, video, and graphics. Multimedia is the stage of computer-based information that allows…
Through the Looking Glass: The Multiple Layers of Multimedia.
ERIC Educational Resources Information Center
D'Ignazio, Fred
1990-01-01
Describes possible future uses of multimedia computers for instructional applications. Highlights include databases; publishing; telecommunications; computers and videocassette recorders (VCRs); audio and video digitizing; video overlay, or genlock; still-image video; videodiscs and CD-ROM; and hypermedia. (LRW)
Good, Daniel W; Delaney, Harry; Laird, Alexander; Hacking, Belinda; Stewart, Grant D; McNeill, S Alan
2016-12-01
The life expectancy of prostate patients is long and patients will spend many years carrying the burdens & benefits of the treatment decisions they have made, therefore, it is vital that decisions on treatments are shared between patient and physician. The objective was to determine if consultation audio-recording improves quality of life, reduces regret or improves patient satisfaction in comparison to standard counselling. In 2012 we initiated consultation audio-recordings, where patients are given a CD of their consultation to keep and replay at home. We conducted a prospective non-randomised study of patient satisfaction, quality of life (QOL) and decision regret at 12 months follow-up using posted validated questionnaires for the audio-recording (AR) patients and a control cohort. Qualitative and thematic analyses were used. Forty of 59 patients in the AR group, and 27 of 45 patients in the control group returned the questionnaires. Patient demographics were similar in both groups with no statistically significant differences between the two groups. Decision regret was lower in the audio-recording group (11/100) vs control group (19/100) (p = 0.04). The risk ratio for not having any long-term decision regret was 5.539 (CI 1.643-18.674), with NNT to prevent regret being 4. Regression analysis showed that receiving audio-recording was strongest predictor for absence of regret even greater than potency and incontinence. The study has shown that audio-recording clinic consultation reduces long-term decision regret, increases patient information recall, understanding and confidence in their decision. There is great potential for further expansion of this low-cost intervention. Copyright © 2014 Royal College of Surgeons of Edinburgh (Scottish charity number SC005317) and Royal College of Surgeons in Ireland. Published by Elsevier Ltd. All rights reserved.
Collusion-Resistant Audio Fingerprinting System in the Modulated Complex Lapped Transform Domain
Garcia-Hernandez, Jose Juan; Feregrino-Uribe, Claudia; Cumplido, Rene
2013-01-01
Collusion-resistant fingerprinting paradigm seems to be a practical solution to the piracy problem as it allows media owners to detect any unauthorized copy and trace it back to the dishonest users. Despite the billionaire losses in the music industry, most of the collusion-resistant fingerprinting systems are devoted to digital images and very few to audio signals. In this paper, state-of-the-art collusion-resistant fingerprinting ideas are extended to audio signals and the corresponding parameters and operation conditions are proposed. Moreover, in order to carry out fingerprint detection using just a fraction of the pirate audio clip, block-based embedding and its corresponding detector is proposed. Extensive simulations show the robustness of the proposed system against average collusion attack. Moreover, by using an efficient Fast Fourier Transform core and standard computer machines it is shown that the proposed system is suitable for real-world scenarios. PMID:23762455
Tape recorder effects on jitter and shimmer extraction.
Doherty, E T; Shipp, T
1988-09-01
To test for possible contamination of acoustic analyses by record/reproduce systems, five sine waves of fixed frequency and amplitude were sampled directly by a computer and recorded simultaneously on four different tape formats (audio and FM reel-to-reel, audio cassette, and video cassette using pulse code modulation). Recordings were digitized on playback and with the direct samples analyzed for fundamental frequency, amplitude, jitter, and shimmer using a zero crossing interpolation scheme. Distortion introduced by any of the data acquisition systems is negligible when extracting average fundamental frequency or average amplitude. For jitter and shimmer estimation, direct sampling or the use of a video cassette recorder with pulse code modulation are clearly superior. FM recorders, although not quite as accurate, provide a satisfactory alternative to those methods. Audio reel-to-reel recordings are marginally adequate for jitter analysis whereas audio cassette recorders can introduce jitter and shimmer values that are greater than some reported values for normal talkers.
Exclusively visual analysis of classroom group interactions
NASA Astrophysics Data System (ADS)
Tucker, Laura; Scherr, Rachel E.; Zickler, Todd; Mazur, Eric
2016-12-01
Large-scale audiovisual data that measure group learning are time consuming to collect and analyze. As an initial step towards scaling qualitative classroom observation, we qualitatively coded classroom video using an established coding scheme with and without its audio cues. We find that interrater reliability is as high when using visual data only—without audio—as when using both visual and audio data to code. Also, interrater reliability is high when comparing use of visual and audio data to visual-only data. We see a small bias to code interactions as group discussion when visual and audio data are used compared with video-only data. This work establishes that meaningful educational observation can be made through visual information alone. Further, it suggests that after initial work to create a coding scheme and validate it in each environment, computer-automated visual coding could drastically increase the breadth of qualitative studies and allow for meaningful educational analysis on a far greater scale.
Kim, Dong-Sun; Kwon, Jin-San
2014-01-01
Research on real-time health systems have received great attention during recent years and the needs of high-quality personal multichannel medical signal compression for personal medical product applications are increasing. The international MPEG-4 audio lossless coding (ALS) standard supports a joint channel-coding scheme for improving compression performance of multichannel signals and it is very efficient compression method for multi-channel biosignals. However, the computational complexity of such a multichannel coding scheme is significantly greater than that of other lossless audio encoders. In this paper, we present a multichannel hardware encoder based on a low-complexity joint-coding technique and shared multiplier scheme for portable devices. A joint-coding decision method and a reference channel selection scheme are modified for a low-complexity joint coder. The proposed joint coding decision method determines the optimized joint-coding operation based on the relationship between the cross correlation of residual signals and the compression ratio. The reference channel selection is designed to select a channel for the entropy coding of the joint coding. The hardware encoder operates at a 40 MHz clock frequency and supports two-channel parallel encoding for the multichannel monitoring system. Experimental results show that the compression ratio increases by 0.06%, whereas the computational complexity decreases by 20.72% compared to the MPEG-4 ALS reference software encoder. In addition, the compression ratio increases by about 11.92%, compared to the single channel based bio-signal lossless data compressor. PMID:25237900
Applications of ENF criterion in forensic audio, video, computer and telecommunication analysis.
Grigoras, Catalin
2007-04-11
This article reports on the electric network frequency criterion as a means of assessing the integrity of digital audio/video evidence and forensic IT and telecommunication analysis. A brief description is given to different ENF types and phenomena that determine ENF variations. In most situations, to reach a non-authenticity opinion, the visual inspection of spectrograms and comparison with an ENF database are enough. A more detailed investigation, in the time domain, requires short time windows measurements and analyses. The stability of the ENF over geographical distances has been established by comparison of synchronized recordings made at different locations on the same network. Real cases are presented, in which the ENF criterion was used to investigate audio and video files created with secret surveillance systems, a digitized audio/video recording and a TV broadcasted reportage. By applying the ENF Criterion in forensic audio/video analysis, one can determine whether and where a digital recording has been edited, establish whether it was made at the time claimed, and identify the time and date of the registering operation.
Effect of Making an Audio Recording of a Term Paper on Writing Quality
ERIC Educational Resources Information Center
Taxis, Tasia M.; Lannin, Amy A.; Selting, Bonita R.; Lamberson, William R.
2014-01-01
Writing-to-learn assignments engage students with a problem while they develop writing skills. It is difficult in large classes to provide training in proofreading and editing techniques. The purpose of this project was to determine if a term paper was improved after making an audio recording of a draft of the paper. Data from 2 years of papers…
ERIC Educational Resources Information Center
Aldalalah, Osamah Ahmad; Fong, Soon Fook
2010-01-01
The purpose of this study was to investigate the effects of modality and redundancy principles on the attitude and learning of music theory among primary pupils of different aptitudes in Jordan. The lesson of music theory was developed in three different modes, audio and image (AI), text with image (TI) and audio with image and text (AIT). The…
The use of ambient audio to increase safety and immersion in location-based games
NASA Astrophysics Data System (ADS)
Kurczak, John Jason
The purpose of this thesis is to propose an alternative type of interface for mobile software being used while walking or running. Our work addresses the problem of visual user interfaces for mobile software be- ing potentially unsafe for pedestrians, and not being very immersive when used for location-based games. In addition, location-based games and applications can be dif- ficult to develop when directly interfacing with the sensors used to track the user's location. These problems need to be addressed because portable computing devices are be- coming a popular tool for navigation, playing games, and accessing the internet while walking. This poses a safety problem for mobile users, who may be paying too much attention to their device to notice and react to hazards in their environment. The difficulty of developing location-based games and other location-aware applications may significantly hinder the prevalence of applications that explore new interaction techniques for ubiquitous computing. We created the TREC toolkit to address the issues with tracking sensors while developing location-based games and applications. We have developed functional location-based applications with TREC to demonstrate the amount of work that can be saved by using this toolkit. In order to have a safer and more immersive alternative to visual interfaces, we have developed ambient audio interfaces for use with mobile applications. Ambient audio uses continuous streams of sound over headphones to present information to mobile users without distracting them from walking safely. In order to test the effectiveness of ambient audio, we ran a study to compare ambient audio with handheld visual interfaces in a location-based game. We compared players' ability to safely navigate the environment, their sense of immersion in the game, and their performance at the in-game tasks. We found that ambient audio was able to significantly increase players' safety and sense of immersion compared to a visual interface, while players performed signifi- cantly better at the game tasks when using the visual interface. This makes ambient audio a legitimate alternative to visual interfaces for mobile users when safety and immersion are a priority.
Human Computer Interface Design Criteria. Volume 1. User Interface Requirements
2010-03-19
Television tuners, including tuner cards for use in computers, shall be equipped with secondary audio program playback circuitry. (c) All training...Shelf CSS Cascading Style Sheets DII Defense Information Infrastructure DISA Defense Information Systems Agency DoD Department of Defense
Sutherland, Rebecca; Trembath, David; Hodge, Antoinette; Drevensek, Suzi; Lee, Sabrena; Silove, Natalie; Roberts, Jacqueline
2017-01-01
Introduction Telehealth can be an effective way to provide speech pathology intervention to children with speech and language impairments. However, the provision of reliable and feasible standardised language assessments via telehealth to establish children's needs for intervention and to monitor progress has not yet been well established. Further, there is limited information about children's reactions to telehealth. This study aimed to examine the reliability and feasibility of conducting standardised language assessment with school-aged children with known or suspected language impairment via a telehealth application using consumer grade computer equipment within a public school setting. Method Twenty-three children (aged 8-12 years) participated. Each child was assessed using a standardised language assessment comprising six subtests. Two subtests were administered by a speech pathologist face-to-face (local clinician) and four subtests were administered via telehealth. All subtests were completed within a single visit to the clinic service, with a break between the face to face and telehealth sessions. The face-to-face clinician completed behaviour observation checklists in the telehealth and face to face conditions and provided feedback on the audio and video quality of the application from the child's point of view. Parent feedback about their child's experience was elicited via survey. Results There was strong inter-rater reliability in the telehealth and face-to-face conditions (correlation coefficients ranged from r = 0.96-1.0 across the subtests) and good agreement on all measures. Similar levels of attention, distractibility and anxiety were observed in the two conditions. Clinicians rated only one session of 23 as having poor audio quality and no sessions were rated as having poor visual quality. Parent and child reactions to the use of telehealth were largely positive and supportive of using telehealth to assess rural children. Discussion The findings support the use of telehealth in the language assessment of school-aged children using a web application and commercially available computer equipment. This reliable and innovative service delivery model has the potential to be used by speech pathologists to provide assessments to children in remote communities.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Falferi, P.; Mezzena, R.; Vitale, S.
1997-08-01
The coupling effects of a commercial dc superconducting quantum interference device (SQUID) to an electrical LC resonator which operates at audio frequencies ({approx}1kHz) with quality factors Q{approx}10{sup 6} are presented. The variations of the resonance frequency of the resonator as functions of the flux applied to the SQUID are due to the SQUID dynamic inductance in good agreement with the predictions of a model. The variations of the quality factor point to a feedback mechanism between the output of the SQUID and the input circuit. {copyright} {ital 1997 American Institute of Physics.}
Point focusing using loudspeaker arrays from the perspective of optimal beamforming.
Bai, Mingsian R; Hsieh, Yu-Hao
2015-06-01
Sound focusing is to create a concentrated acoustic field in the region surrounded by a loudspeaker array. This problem was tackled in the previous research via the Helmholtz integral approach, brightness control, acoustic contrast control, etc. In this paper, the same problem was revisited from the perspective of beamforming. A source array model is reformulated in terms of the steering matrix between the source and the field points, which lends itself to the use of beamforming algorithms such as minimum variance distortionless response (MVDR) and linearly constrained minimum variance (LCMV) originally intended for sensor arrays. The beamforming methods are compared with the conventional methods in terms of beam pattern, directional index, and control effort. Objective tests are conducted to assess the audio quality by using perceptual evaluation of audio quality (PEAQ). Experiments of produced sound field and listening tests are conducted in a listening room, with results processed using analysis of variance and regression analysis. In contrast to the conventional energy-based methods, the results have shown that the proposed methods are phase-sensitive in light of the distortionless constraint in formulating the array filters, which helps enhance audio quality and focusing performance.
Eyben, Florian; Weninger, Felix; Lehment, Nicolas; Schuller, Björn; Rigoll, Gerhard
2013-01-01
Without doubt general video and sound, as found in large multimedia archives, carry emotional information. Thus, audio and video retrieval by certain emotional categories or dimensions could play a central role for tomorrow's intelligent systems, enabling search for movies with a particular mood, computer aided scene and sound design in order to elicit certain emotions in the audience, etc. Yet, the lion's share of research in affective computing is exclusively focusing on signals conveyed by humans, such as affective speech. Uniting the fields of multimedia retrieval and affective computing is believed to lend to a multiplicity of interesting retrieval applications, and at the same time to benefit affective computing research, by moving its methodology "out of the lab" to real-world, diverse data. In this contribution, we address the problem of finding "disturbing" scenes in movies, a scenario that is highly relevant for computer-aided parental guidance. We apply large-scale segmental feature extraction combined with audio-visual classification to the particular task of detecting violence. Our system performs fully data-driven analysis including automatic segmentation. We evaluate the system in terms of mean average precision (MAP) on the official data set of the MediaEval 2012 evaluation campaign's Affect Task, which consists of 18 original Hollywood movies, achieving up to .398 MAP on unseen test data in full realism. An in-depth analysis of the worth of individual features with respect to the target class and the system errors is carried out and reveals the importance of peak-related audio feature extraction and low-level histogram-based video analysis.
Eyben, Florian; Weninger, Felix; Lehment, Nicolas; Schuller, Björn; Rigoll, Gerhard
2013-01-01
Without doubt general video and sound, as found in large multimedia archives, carry emotional information. Thus, audio and video retrieval by certain emotional categories or dimensions could play a central role for tomorrow's intelligent systems, enabling search for movies with a particular mood, computer aided scene and sound design in order to elicit certain emotions in the audience, etc. Yet, the lion's share of research in affective computing is exclusively focusing on signals conveyed by humans, such as affective speech. Uniting the fields of multimedia retrieval and affective computing is believed to lend to a multiplicity of interesting retrieval applications, and at the same time to benefit affective computing research, by moving its methodology “out of the lab” to real-world, diverse data. In this contribution, we address the problem of finding “disturbing” scenes in movies, a scenario that is highly relevant for computer-aided parental guidance. We apply large-scale segmental feature extraction combined with audio-visual classification to the particular task of detecting violence. Our system performs fully data-driven analysis including automatic segmentation. We evaluate the system in terms of mean average precision (MAP) on the official data set of the MediaEval 2012 evaluation campaign's Affect Task, which consists of 18 original Hollywood movies, achieving up to .398 MAP on unseen test data in full realism. An in-depth analysis of the worth of individual features with respect to the target class and the system errors is carried out and reveals the importance of peak-related audio feature extraction and low-level histogram-based video analysis. PMID:24391704
Evaluation of architectures for an ASP MPEG-4 decoder using a system-level design methodology
NASA Astrophysics Data System (ADS)
Garcia, Luz; Reyes, Victor; Barreto, Dacil; Marrero, Gustavo; Bautista, Tomas; Nunez, Antonio
2005-06-01
Trends in multimedia consumer electronics, digital video and audio, aim to reach users through low-cost mobile devices connected to data broadcasting networks with limited bandwidth. An emergent broadcasting network is the digital audio broadcasting network (DAB) which provides CD quality audio transmission together with robustness and efficiency techniques to allow good quality reception in motion conditions. This paper focuses on the system-level evaluation of different architectural options to allow low bandwidth digital video reception over DAB, based on video compression techniques. Profiling and design space exploration techniques are applied over the ASP MPEG-4 decoder in order to find out the best HW/SW partition given the application and platform constraints. An innovative SystemC-based system-level design tool, called CASSE, is being used for modelling, exploration and evaluation of different ASP MPEG-4 decoder HW/SW partitions. System-level trade offs and quantitative data derived from this analysis are also presented in this work.
A Methodology and Analysis for Cost-Effective Training in the AN/TSQ-73 Missile Minder
1978-02-01
subsequent users must join the program in progress. 10. Languae Laboratory - Audio, Active - Compare Mode - Al audio presenta- tional device that distributes...initial performanceof the Fystem, change inputs to or elements within the system and 4note changes in the performance of the system. 33. Teaching...Any contest, governed by rules, between teams or individuals, where the contest is a dynamic model of some real system, and a computer is used in
ERIC Educational Resources Information Center
Smith, Glenn Gordon
2012-01-01
This study compared books with embedded computer games (via pentop computers with microdot paper and audio feedback) with regular books with maps, in terms of fifth graders' comprehension and retention of spatial details from stories. One group read a story in hard copy with embedded computer games, the other group read it in regular book format…
Soprano and source: A laryngographic analysis
NASA Astrophysics Data System (ADS)
Bateman, Laura Anne
2005-04-01
Popular music in the 21st century uses a particular singing quality for female voice that is quite different from the trained classical singing quality. Classical quality has been the subject of a vast body of research, whereas research that deals with non-classical qualities is limited. In order to learn more about these issues, the author chose to do research on singing qualities using a variety of standard voice quality tests. This paper looks at voice qualities found in various different styles of singing: Classical, Belt, Legit, R&B, Jazz, Country, and Pop. The data was elicited from a professional soprano and the voice qualities reflect industry standards. The data set for this paper is limited to samples using the vowel [i]. Laryngographic (LGG) data was generated simultaneously with the audio samples. This paper will focus on the results of the LGG analysis; however, an audio analysis was also performed using Spectrogram, LPC, and FFT. Data from the LGG is used to calculate the contact quotient, speed quotient, and ascending slope. The LGG waveform is also visually assessed. The LGG analysis gives insights into the source vibration for the different singing styles.
Computer Series, 86. Bits and Pieces, 35.
ERIC Educational Resources Information Center
Moore, John W., Ed.
1987-01-01
Describes eight applications of the use of computers in teaching chemistry. Includes discussions of audio frequency measurements of heat capacity ratios, quantum mechanics, ab initio calculations, problem solving using spreadsheets, simplex optimization, faradaic impedance diagrams, and the recording and tabulation of student laboratory data. (TW)
NASA Astrophysics Data System (ADS)
Li, Wei; Xiao, Chuan; Liu, Yaduo
2013-12-01
Audio identification via fingerprint has been an active research field for years. However, most previously reported methods work on the raw audio format in spite of the fact that nowadays compressed format audio, especially MP3 music, has grown into the dominant way to store music on personal computers and/or transmit it over the Internet. It will be interesting if a compressed unknown audio fragment could be directly recognized from the database without decompressing it into the wave format at first. So far, very few algorithms run directly on the compressed domain for music information retrieval, and most of them take advantage of the modified discrete cosine transform coefficients or derived cepstrum and energy type of features. As a first attempt, we propose in this paper utilizing compressed domain auditory Zernike moment adapted from image processing techniques as the key feature to devise a novel robust audio identification algorithm. Such fingerprint exhibits strong robustness, due to its statistically stable nature, against various audio signal distortions such as recompression, noise contamination, echo adding, equalization, band-pass filtering, pitch shifting, and slight time scale modification. Experimental results show that in a music database which is composed of 21,185 MP3 songs, a 10-s long music segment is able to identify its original near-duplicate recording, with average top-5 hit rate up to 90% or above even under severe audio signal distortions.
Implementation of audio computer-assisted interviewing software in HIV/AIDS research.
Pluhar, Erika; McDonnell Holstad, Marcia; Yeager, Katherine A; Denzmore-Nwagbara, Pamela; Corkran, Carol; Fielder, Bridget; McCarty, Frances; Diiorio, Colleen
2007-01-01
Computer-assisted interviewing (CAI) has begun to play a more prominent role in HIV/AIDS prevention research. Despite the increased popularity of CAI, particularly audio computer-assisted self-interviewing (ACASI), some research teams are still reluctant to implement ACASI technology because of lack of familiarity with the practical issues related to using these software packages. The purpose of this report is to describe the implementation of one particular ACASI software package, the Questionnaire Development System (QDS; Nova Research Company, Bethesda, MD), in several nursing and HIV/AIDS prevention research settings. The authors present acceptability and satisfaction data from two large-scale public health studies in which they have used QDS with diverse populations. They also address issues related to developing and programming a questionnaire; discuss practical strategies related to planning for and implementing ACASI in the field, including selecting equipment, training staff, and collecting and transferring data; and summarize advantages and disadvantages of computer-assisted research methods.
Audio-visual presentation of information for informed consent for participation in clinical trials.
Synnot, Anneliese; Ryan, Rebecca; Prictor, Megan; Fetherstonhaugh, Deirdre; Parker, Barbara
2014-05-09
Informed consent is a critical component of clinical research. Different methods of presenting information to potential participants of clinical trials may improve the informed consent process. Audio-visual interventions (presented, for example, on the Internet or on DVD) are one such method. We updated a 2008 review of the effects of these interventions for informed consent for trial participation. To assess the effects of audio-visual information interventions regarding informed consent compared with standard information or placebo audio-visual interventions regarding informed consent for potential clinical trial participants, in terms of their understanding, satisfaction, willingness to participate, and anxiety or other psychological distress. We searched: the Cochrane Central Register of Controlled Trials (CENTRAL), The Cochrane Library, issue 6, 2012; MEDLINE (OvidSP) (1946 to 13 June 2012); EMBASE (OvidSP) (1947 to 12 June 2012); PsycINFO (OvidSP) (1806 to June week 1 2012); CINAHL (EbscoHOST) (1981 to 27 June 2012); Current Contents (OvidSP) (1993 Week 27 to 2012 Week 26); and ERIC (Proquest) (searched 27 June 2012). We also searched reference lists of included studies and relevant review articles, and contacted study authors and experts. There were no language restrictions. We included randomised and quasi-randomised controlled trials comparing audio-visual information alone, or in conjunction with standard forms of information provision (such as written or verbal information), with standard forms of information provision or placebo audio-visual information, in the informed consent process for clinical trials. Trials involved individuals or their guardians asked to consider participating in a real or hypothetical clinical study. (In the earlier version of this review we only included studies evaluating informed consent interventions for real studies). Two authors independently assessed studies for inclusion and extracted data. We synthesised the findings using meta-analysis, where possible, and narrative synthesis of results. We assessed the risk of bias of individual studies and considered the impact of the quality of the overall evidence on the strength of the results. We included 16 studies involving data from 1884 participants. Nine studies included participants considering real clinical trials, and eight included participants considering hypothetical clinical trials, with one including both. All studies were conducted in high-income countries.There is still much uncertainty about the effect of audio-visual informed consent interventions on a range of patient outcomes. However, when considered across comparisons, we found low to very low quality evidence that such interventions may slightly improve knowledge or understanding of the parent trial, but may make little or no difference to rate of participation or willingness to participate. Audio-visual presentation of informed consent may improve participant satisfaction with the consent information provided. However its effect on satisfaction with other aspects of the process is not clear. There is insufficient evidence to draw conclusions about anxiety arising from audio-visual informed consent. We found conflicting, very low quality evidence about whether audio-visual interventions took more or less time to administer. No study measured researcher satisfaction with the informed consent process, nor ease of use.The evidence from real clinical trials was rated as low quality for most outcomes, and for hypothetical studies, very low. We note, however, that this was in large part due to poor study reporting, the hypothetical nature of some studies and low participant numbers, rather than inconsistent results between studies or confirmed poor trial quality. We do not believe that any studies were funded by organisations with a vested interest in the results. The value of audio-visual interventions as a tool for helping to enhance the informed consent process for people considering participating in clinical trials remains largely unclear, although trends are emerging with regard to improvements in knowledge and satisfaction. Many relevant outcomes have not been evaluated in randomised trials. Triallists should continue to explore innovative methods of providing information to potential trial participants during the informed consent process, mindful of the range of outcomes that the intervention should be designed to achieve, and balancing the resource implications of intervention development and delivery against the purported benefits of any intervention.More trials, adhering to CONSORT standards, and conducted in settings and populations underserved in this review, i.e. low- and middle-income countries and people with low literacy, would strengthen the results of this review and broaden its applicability. Assessing process measures, such as time taken to administer the intervention and researcher satisfaction, would inform the implementation of audio-visual consent materials.
A Proposal on the Validation Model of Equivalence between PBLT and CBLT
ERIC Educational Resources Information Center
Chen, Huilin
2014-01-01
The validity of the computer-based language test is possibly affected by three factors: computer familiarity, audio-visual cognitive competence, and other discrepancies in construct. Therefore, validating the equivalence between the paper-and-pencil language test and the computer-based language test is a key step in the procedure of designing a…
Audio-visual presentation of information for informed consent for participation in clinical trials.
Ryan, R E; Prictor, M J; McLaughlin, K J; Hill, S J
2008-01-23
Informed consent is a critical component of clinical research. Different methods of presenting information to potential participants of clinical trials may improve the informed consent process. Audio-visual interventions (presented for example on the Internet, DVD, or video cassette) are one such method. To assess the effects of providing audio-visual information alone, or in conjunction with standard forms of information provision, to potential clinical trial participants in the informed consent process, in terms of their satisfaction, understanding and recall of information about the study, level of anxiety and their decision whether or not to participate. We searched: the Cochrane Consumers and Communication Review Group Specialised Register (searched 20 June 2006); the Cochrane Central Register of Controlled Trials (CENTRAL), The Cochrane Library, issue 2, 2006; MEDLINE (Ovid) (1966 to June week 1 2006); EMBASE (Ovid) (1988 to 2006 week 24); and other databases. We also searched reference lists of included studies and relevant review articles, and contacted study authors and experts. There were no language restrictions. Randomised and quasi-randomised controlled trials comparing audio-visual information alone, or in conjunction with standard forms of information provision (such as written or oral information as usually employed in the particular service setting), with standard forms of information provision alone, in the informed consent process for clinical trials. Trials involved individuals or their guardians asked to participate in a real (not hypothetical) clinical study. Two authors independently assessed studies for inclusion and extracted data. Due to heterogeneity no meta-analysis was possible; we present the findings in a narrative review. We included 4 trials involving data from 511 people. Studies were set in the USA and Canada. Three were randomised controlled trials (RCTs) and the fourth a quasi-randomised trial. Their quality was mixed and results should be interpreted with caution. Considerable uncertainty remains about the effects of audio-visual interventions, compared with standard forms of information provision (such as written or oral information normally used in the particular setting), for use in the process of obtaining informed consent for clinical trials. Audio-visual interventions did not consistently increase participants' levels of knowledge/understanding (assessed in four studies), although one study showed better retention of knowledge amongst intervention recipients. An audio-visual intervention may transiently increase people's willingness to participate in trials (one study), but this was not sustained at two to four weeks post-intervention. Perceived worth of the trial did not appear to be influenced by an audio-visual intervention (one study), but another study suggested that the quality of information disclosed may be enhanced by an audio-visual intervention. Many relevant outcomes including harms were not measured. The heterogeneity in results may reflect the differences in intervention design, content and delivery, the populations studied and the diverse methods of outcome assessment in included studies. The value of audio-visual interventions for people considering participating in clinical trials remains unclear. Evidence is mixed as to whether audio-visual interventions enhance people's knowledge of the trial they are considering entering, and/or the health condition the trial is designed to address; one study showed improved retention of knowledge amongst intervention recipients. The intervention may also have small positive effects on the quality of information disclosed, and may increase willingness to participate in the short-term; however the evidence is weak. There were no data for several primary outcomes, including harms. In the absence of clear results, triallists should continue to explore innovative methods of providing information to potential trial participants. Further research should take the form of high-quality randomised controlled trials, with clear reporting of methods. Studies should conduct content assessment of audio-visual and other innovative interventions for people of differing levels of understanding and education; also for different age and cultural groups. Researchers should assess systematically the effects of different intervention components and delivery characteristics, and should involve consumers in intervention development. Studies should assess additional outcomes relevant to individuals' decisional capacity, using validated tools, including satisfaction; anxiety; and adherence to the subsequent trial protocol.
Using speech recognition to enhance the Tongue Drive System functionality in computer access.
Huo, Xueliang; Ghovanloo, Maysam
2011-01-01
Tongue Drive System (TDS) is a wireless tongue operated assistive technology (AT), which can enable people with severe physical disabilities to access computers and drive powered wheelchairs using their volitional tongue movements. TDS offers six discrete commands, simultaneously available to the users, for pointing and typing as a substitute for mouse and keyboard in computer access, respectively. To enhance the TDS performance in typing, we have added a microphone, an audio codec, and a wireless audio link to its readily available 3-axial magnetic sensor array, and combined it with a commercially available speech recognition software, the Dragon Naturally Speaking, which is regarded as one of the most efficient ways for text entry. Our preliminary evaluations indicate that the combined TDS and speech recognition technologies can provide end users with significantly higher performance than using each technology alone, particularly in completing tasks that require both pointing and text entry, such as web surfing.
Merchant, Raina M; Abella, Benjamin S; Abotsi, Edem J; Smith, Thomas M; Long, Judith A; Trudeau, Martha E; Leary, Marion; Groeneveld, Peter W; Becker, Lance B; Asch, David A
2010-06-01
Given the ubiquitous presence of cellular telephones, we seek to evaluate the extent to which prerecorded audio cardiopulmonary resuscitation (CPR) instructions delivered by a cell telephone will improve the quality of CPR provided by untrained and trained lay rescuers. We randomly assigned both previously CPR trained and untrained volunteers to perform CPR on a manikin for 3 minutes with or without audio assistance from a cell telephone programmed to provide CPR instructions. We measured CPR quality metrics-pauses (ie, no flow time), compression rate (minute), depth (millimeters), and hand placement (percentage correct)-across the 4 groups defined by being either CPR trained or untrained and receiving or not receiving cell telephone CPR instructions. There was no difference in CPR measures for participants who had or had not received previous CPR training. Participants using the cell telephone aid performed better compression rate (100/minute [95% confidence interval (CI) 97 to 103/minute] versus 44/minute [95% CI 38 to 50/minute]), compression depth (41 mm [95% CI 38 to 44 mm] versus 31 mm [95% CI 28 to 34 mm]), hand placement (97% [95% CI 94% to 100%] versus 75% [95% CI 68% to 83%] correct), and fewer pauses (74 seconds [95% CI 72 to 76 seconds] versus 89 seconds [95% CI 80 to 98 seconds]) compared with participants without the cell telephone aid. A simple audio program that can be made available for cell telephones increases the quality of bystander CPR in a manikin simulation. Copyright (c) 2009 American College of Emergency Physicians. Published by Mosby, Inc. All rights reserved.
Multifunctional microcontrollable interface module
NASA Astrophysics Data System (ADS)
Spitzer, Mark B.; Zavracky, Paul M.; Rensing, Noa M.; Crawford, J.; Hockman, Angela H.; Aquilino, P. D.; Girolamo, Henry J.
2001-08-01
This paper reports the development of a complete eyeglass- mounted computer interface system including display, camera and audio subsystems. The display system provides an SVGA image with a 20 degree horizontal field of view. The camera system has been optimized for face recognition and provides a 19 degree horizontal field of view. A microphone and built-in pre-amp optimized for voice recognition and a speaker on an articulated arm are included for audio. An important feature of the system is a high degree of adjustability and reconfigurability. The system has been developed for testing by the Military Police, in a complete system comprising the eyeglass-mounted interface, a wearable computer, and an RF link. Details of the design, construction, and performance of the eyeglass-based system are discussed.
Aviation & Space Education: A Teacher's Resource Guide.
ERIC Educational Resources Information Center
Texas State Dept. of Aviation, Austin.
This resource guide contains information on curriculum guides, resources for teachers, computer software and computer related programs, audio/visual presentations, model aircraft and demonstration aids, training seminars and career education, and an aerospace bibliography for primary grades. Each entry includes all or some of the following items:…
Savari, Maryam; Abdul Wahab, Ainuddin Wahid; Anuar, Nor Badrul
2016-09-01
Audio forgery is any act of tampering, illegal copy and fake quality in the audio in a criminal way. In the last decade, there has been increasing attention to the audio forgery detection due to a significant increase in the number of forge in different type of audio. There are a number of methods for forgery detection, which electric network frequency (ENF) is one of the powerful methods in this area for forgery detection in terms of accuracy. In spite of suitable accuracy of ENF in a majority of plug-in powered devices, the weak accuracy of ENF in audio forgery detection for battery-powered devices, especially in laptop and mobile phone, can be consider as one of the main obstacles of the ENF. To solve the ENF problem in terms of accuracy in battery-powered devices, a combination method of ENF and phase feature is proposed. From experiment conducted, ENF alone give 50% and 60% accuracy for forgery detection in mobile phone and laptop respectively, while the proposed method shows 88% and 92% accuracy respectively, for forgery detection in battery-powered devices. The results lead to higher accuracy for forgery detection with the combination of ENF and phase feature. Copyright © 2016 Elsevier Ireland Ltd. All rights reserved.
Incorporating Auditory Models in Speech/Audio Applications
NASA Astrophysics Data System (ADS)
Krishnamoorthi, Harish
2011-12-01
Following the success in incorporating perceptual models in audio coding algorithms, their application in other speech/audio processing systems is expanding. In general, all perceptual speech/audio processing algorithms involve minimization of an objective function that directly/indirectly incorporates properties of human perception. This dissertation primarily investigates the problems associated with directly embedding an auditory model in the objective function formulation and proposes possible solutions to overcome high complexity issues for use in real-time speech/audio algorithms. Specific problems addressed in this dissertation include: 1) the development of approximate but computationally efficient auditory model implementations that are consistent with the principles of psychoacoustics, 2) the development of a mapping scheme that allows synthesizing a time/frequency domain representation from its equivalent auditory model output. The first problem is aimed at addressing the high computational complexity involved in solving perceptual objective functions that require repeated application of auditory model for evaluation of different candidate solutions. In this dissertation, a frequency pruning and a detector pruning algorithm is developed that efficiently implements the various auditory model stages. The performance of the pruned model is compared to that of the original auditory model for different types of test signals in the SQAM database. Experimental results indicate only a 4-7% relative error in loudness while attaining up to 80-90 % reduction in computational complexity. Similarly, a hybrid algorithm is developed specifically for use with sinusoidal signals and employs the proposed auditory pattern combining technique together with a look-up table to store representative auditory patterns. The second problem obtains an estimate of the auditory representation that minimizes a perceptual objective function and transforms the auditory pattern back to its equivalent time/frequency representation. This avoids the repeated application of auditory model stages to test different candidate time/frequency vectors in minimizing perceptual objective functions. In this dissertation, a constrained mapping scheme is developed by linearizing certain auditory model stages that ensures obtaining a time/frequency mapping corresponding to the estimated auditory representation. This paradigm was successfully incorporated in a perceptual speech enhancement algorithm and a sinusoidal component selection task.
Astronaut James Newman works with computers and GPS
1993-09-20
STS051-16-028 (12-22 Sept 1993) --- On Discovery's middeck, astronaut James H. Newman, mission specialist, works with an array of computers, including one devoted to Global Positioning System (GPS) operations, a general portable onboard computer displaying a tracking map, a portable audio data modem and another payload and general support computer. Newman was joined by four other NASA astronauts for almost ten full days in space.
A Pilot Study of a Self-Voicing Computer Program for Prealgebra Math Problems
ERIC Educational Resources Information Center
Beal, Carole R.; Rosenblum, L. Penny; Smith, Derrick W.
2011-01-01
Fourteen students with visual impairments in Grades 5-12 participated in the field-testing of AnimalWatch-VI-Beta. This computer program delivered 12 prealgebra math problems and hints through a self-voicing audio feature. The students provided feedback about how the computer program can be improved and expanded to make it accessible to all users.…
Free Oscilloscope Web App Using a Computer Mic, Built-In Sound Library, or Your Own Files
ERIC Educational Resources Information Center
Ball, Edward; Ruiz, Frances; Ruiz, Michael J.
2017-01-01
We have developed an online oscilloscope program which allows users to see waveforms by utilizing their computer microphones, selecting from our library of over 30 audio files, and opening any *.mp3 or *.wav file on their computers. The oscilloscope displays real-time signals against time. The oscilloscope has been calibrated so one can make…
Challenges in Transcribing Multimodal Data: A Case Study
ERIC Educational Resources Information Center
Helm, Francesca; Dooly, Melinda
2017-01-01
Computer-mediated communication (CMC) once meant principally text-based communication mediated by computers, but rapid technological advances in recent years have heralded an era of multimodal communication with a growing emphasis on audio and video synchronous interaction. As CMC, in all its variants (text chats, video chats, forums, blogs, SMS,…
Sixth Annual Conference on Computers, Freedom, and Privacy: The RealAudio Proceedings.
ERIC Educational Resources Information Center
Glover, Barbara; Meernik, Mary
1996-01-01
Reviews the sixth Conference on Computers, Freedom, and Privacy (CFP) held in March 1996. Highlights include the Communications Decency Act, part of the 1996 Telecommunications Reform Act; European views; Internet service providers; limiting online speech on campus; cryptography; the global information infrastructure; copyright; and China and the…
How Forcefully Should Universities Enforce Copyright Law on Audio Files?
ERIC Educational Resources Information Center
McCollum, Kelly
1999-01-01
The Recording Industry Association of America is aggressively pursuing copyright violations on campuses concerning MP3 music recordings being exchanged on computer networks. Carnegie Mellon University (Pennsylvania), to avoid litigation, has been searching public folders of students' computers to find illegally copied MP3s. Controversy over…
Optical Fiber In The Loop: Features And Applications
NASA Astrophysics Data System (ADS)
Shariati, Ross
1986-01-01
It is expected that there would be various demands for digital capacity, from a few kilobits per second for such services as facsimile, data entry, and provision of audio and graphic for teleconferencing, to about 56Kb/sec for electronic mail and integrated work stations, and higher speeds for cable television, high resolution TV, and computer-aided engineering. Fiber optics has been proven-in from an economic standpoint to provide the above-mentioned services. This is primarily due to the fact that in less than five years optical line rates have leaped from 45Mb/s to gigabit rates, therefore reducing the cost per DS3 of capacity, and the price of high quality fiber cable has taken a nosedive.
Creating accessible science museums with user-activated environmental audio beacons (ping!).
Landau, Steven; Wiener, William; Naghshineh, Koorosh; Giusti, Ellen
2005-01-01
In 2003, Touch Graphics Company carried out research on a new invention that promises to improve accessibility to science museums for visitors who are visually impaired. The system, nicknamed Ping!, allows users to navigate an exhibit area, listen to audio descriptions, and interact with exhibits using a cell phone-based interface. The system relies on computer telephony, and it incorporates a network of wireless environmental audio beacons that can be triggered by users wishing to travel to destinations they choose. User testing indicates that the system is effective, both as a way-finding tool and as a means of providing accessible information on museum content. Follow-up development projects will determine if this approach can be successfully implemented in other settings and for other user populations.
Effect of Spinal Manipulative Therapy on the Singing Voice.
Fachinatto, Ana Paula A; Duprat, André de Campos; Silva, Marta Andrada E; Bracher, Eduardo Sawaya Botelho; Benedicto, Camila de Carvalho; Luz, Victor Botta Colangelo; Nogueira, Maruan Nogueira; Fonseca, Beatriz Suster Gomes
2015-09-01
This study investigated the effect of spinal manipulative therapy (SMT) on the singing voice of male individuals. Randomized, controlled, case-crossover trial. Twenty-nine subjects were selected among male members of the Heralds of the Gospel. This association was chosen because it is a group of persons with similar singing activities. Participants were randomly assigned to two groups: (A) chiropractic SMT procedure and (B) nontherapeutic transcutaneous electrical nerve stimulation (TENS) procedure. Recordings of the singing voice of each participant were taken immediately before and after the procedures. After a 14-day period, procedures were switched between groups: participants who underwent SMT on the first day were subjected to TENS and vice versa. Recordings were subjected to perceptual audio and acoustic evaluations. The same recording segment of each participant was selected. Perceptual audio evaluation was performed by a specialist panel (SP). Recordings of each participant were randomly presented thus making the SP blind to intervention type and recording session (before/after intervention). Recordings compiled in a randomized order were also subjected to acoustic evaluation. No differences in the quality of the singing on perceptual audio evaluation were observed between TENS and SMT. No differences in the quality of the singing voice of asymptomatic male singers were observed on perceptual audio evaluation or acoustic evaluation after a single spinal manipulative intervention of the thoracic and cervical spine. Copyright © 2015 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Wofford, James L; Campos, Claudia L; Johnson, Dominic A; Brown, Monica T
2012-01-01
The advent of more mobile, more reliable, and more affordable videoconferencing technology finally makes it realistic to offer remote foreign language interpretation in the office setting. Still, such technologies deserve proof of acceptability to clinicians and patients before there is widespread acceptance and routine use. We sought to examine: (1) the audio and video technical fidelity of iPad/Facetime(TM) software, (2) the acceptability of videoconferencing to patients and clinicians. The convenience sample included Spanish-speaking adult patients at a community health care medicine clinic in 2011. Videoconferencing was conducted using two iPads(TM) connecting patient/physician located in the clinic examination room, and the interpreter in a remote/separate office in the same building. A five-item survey was used to solicit opinions on overall quality of the videoconferencing device, audio/video integrity/fidelity, perception of encounter duration, and attitude toward future use. Twenty-five patients, 18 clinicians and 5 interpreters participated in the project. Most patients (24/25) rated overall quality of videoconferencing as good/excellent with only 1 'fair' rating. Eleven patients rated the amount of time as no longer than in-person, and nine reported it as shorter than inperson. Most patients, 94.0% (24/25), favoured using videoconferencing during future visits. For the 18 clinicians, the results were similar. Based on our experience at a singlesite community health centre, the videoconferencing technology appeared to be flawless, and both patients and clinicians were satisfied. Expansion of videoconferencing to other off-site healthcare professionals should be considered in the search for more cost-effective healthcare.
Wilbiks, Jonathan M P; Dyson, Benjamin J
2016-01-01
Over 5 experiments, we challenge the idea that the capacity of audio-visual integration need be fixed at 1 item. We observe that the conditions under which audio-visual integration is most likely to exceed 1 occur when stimulus change operates at a slow rather than fast rate of presentation and when the task is of intermediate difficulty such as when low levels of proactive interference (3 rather than 8 interfering visual presentations) are combined with the temporal unpredictability of the critical frame (Experiment 2), or, high levels of proactive interference are combined with the temporal predictability of the critical frame (Experiment 4). Neural data suggest that capacity might also be determined by the quality of perceptual information entering working memory. Experiment 5 supported the proposition that audio-visual integration was at play during the previous experiments. The data are consistent with the dynamic nature usually associated with cross-modal binding, and while audio-visual integration capacity likely cannot exceed uni-modal capacity estimates, performance may be better than being able to associate only one visual stimulus with one auditory stimulus.
Wilbiks, Jonathan M. P.; Dyson, Benjamin J.
2016-01-01
Over 5 experiments, we challenge the idea that the capacity of audio-visual integration need be fixed at 1 item. We observe that the conditions under which audio-visual integration is most likely to exceed 1 occur when stimulus change operates at a slow rather than fast rate of presentation and when the task is of intermediate difficulty such as when low levels of proactive interference (3 rather than 8 interfering visual presentations) are combined with the temporal unpredictability of the critical frame (Experiment 2), or, high levels of proactive interference are combined with the temporal predictability of the critical frame (Experiment 4). Neural data suggest that capacity might also be determined by the quality of perceptual information entering working memory. Experiment 5 supported the proposition that audio-visual integration was at play during the previous experiments. The data are consistent with the dynamic nature usually associated with cross-modal binding, and while audio-visual integration capacity likely cannot exceed uni-modal capacity estimates, performance may be better than being able to associate only one visual stimulus with one auditory stimulus. PMID:27977790
Reconstruction of audio waveforms from spike trains of artificial cochlea models
Zai, Anja T.; Bhargava, Saurabh; Mesgarani, Nima; Liu, Shih-Chii
2015-01-01
Spiking cochlea models describe the analog processing and spike generation process within the biological cochlea. Reconstructing the audio input from the artificial cochlea spikes is therefore useful for understanding the fidelity of the information preserved in the spikes. The reconstruction process is challenging particularly for spikes from the mixed signal (analog/digital) integrated circuit (IC) cochleas because of multiple non-linearities in the model and the additional variance caused by random transistor mismatch. This work proposes an offline method for reconstructing the audio input from spike responses of both a particular spike-based hardware model called the AEREAR2 cochlea and an equivalent software cochlea model. This method was previously used to reconstruct the auditory stimulus based on the peri-stimulus histogram of spike responses recorded in the ferret auditory cortex. The reconstructed audio from the hardware cochlea is evaluated against an analogous software model using objective measures of speech quality and intelligibility; and further tested in a word recognition task. The reconstructed audio under low signal-to-noise (SNR) conditions (SNR < –5 dB) gives a better classification performance than the original SNR input in this word recognition task. PMID:26528113
ERIC Educational Resources Information Center
Bryce, C. F. A.; Stewart, A. M.
A brief review of the characteristics of computer assisted instruction and the attributes of audiovisual media introduces this report on a project designed to improve the effectiveness of computer assisted learning through the incorporation of audiovisual materials. A discussion of the implications of research findings on the design and layout of…
Video mining using combinations of unsupervised and supervised learning techniques
NASA Astrophysics Data System (ADS)
Divakaran, Ajay; Miyahara, Koji; Peker, Kadir A.; Radhakrishnan, Regunathan; Xiong, Ziyou
2003-12-01
We discuss the meaning and significance of the video mining problem, and present our work on some aspects of video mining. A simple definition of video mining is unsupervised discovery of patterns in audio-visual content. Such purely unsupervised discovery is readily applicable to video surveillance as well as to consumer video browsing applications. We interpret video mining as content-adaptive or "blind" content processing, in which the first stage is content characterization and the second stage is event discovery based on the characterization obtained in stage 1. We discuss the target applications and find that using a purely unsupervised approach are too computationally complex to be implemented on our product platform. We then describe various combinations of unsupervised and supervised learning techniques that help discover patterns that are useful to the end-user of the application. We target consumer video browsing applications such as commercial message detection, sports highlights extraction etc. We employ both audio and video features. We find that supervised audio classification combined with unsupervised unusual event discovery enables accurate supervised detection of desired events. Our techniques are computationally simple and robust to common variations in production styles etc.
Howarth, Ana; Perkins-Porras, Linda; Smith, Jared G; Subramaniam, Jeevakan; Copland, Claire; Hurley, Mike; Beith, Iain; Riaz, Muhammad; Ussher, Michael
2016-06-02
The burden of chronic pain is a major challenge, impacting the quality of life of patients. Intensive programmes of mindfulness-based therapy can help patients to cope with chronic pain but can be time consuming and require a trained specialist to implement. The self-management model of care is now integral to the care of patients with chronic pain; home-based interventions can be very acceptable, making a compelling argument for investigating brief, self-management interventions. The aim of this study is two-fold: to assess the immediate effects of a brief self-help mindfulness intervention for coping with chronic pain and to assess the feasibility of conducting a definitive randomized controlled trial to determine the effectiveness of such an intervention. A randomized controlled pilot study will be conducted to evaluate a brief mindfulness intervention for those with chronic pain. Ninety chronic pain patients who attend hospital outpatient clinics will be recruited and allocated randomly to either the control or treatment group on a 1:1 basis using the computer-generated list of random numbers. The treatment group receives mindfulness audios and the control group receives audios of readings from a non-fiction book, all of which are 15 minutes in length. Immediate effects of the intervention are assessed with brief psychological measures immediately before and after audio use. Mindfulness, mood, health-related quality of life, pain catastrophizing and experience of the intervention are assessed with standardized measures, brief ratings and brief telephone follow-ups, at baseline and after one week and one month. Feasibility is assessed by estimation of effect sizes for outcomes, patient adherence and experience, and appraisal of resource allocation in provision of the intervention. This trial will assess whether a brief mindfulness-based intervention is effective for immediately reducing perceived distress and pain with the side effect of increasing relaxation in chronic pain patients and will determine the feasibility of conducting a definitive randomized controlled trial. Patient recruitment began in January 2015 and is due to be completed in June 2016. ISRCTN61538090 Registered 20 April 2015.
Real-time implementation of second generation of audio multilevel information coding
NASA Astrophysics Data System (ADS)
Ali, Murtaza; Tewfik, Ahmed H.; Viswanathan, V.
1994-03-01
This paper describes real-time implementation of a novel wavelet- based audio compression method. This method is based on the discrete wavelet (DWT) representation of signals. A bit allocation procedure is used to allocate bits to the transform coefficients in an adaptive fashion. The bit allocation procedure has been designed to take advantage of the masking effect in human hearing. The procedure minimizes the number of bits required to represent each frame of audio signals at a fixed distortion level. The real-time implementation provides almost transparent compression of monophonic CD quality audio signals (samples at 44.1 KHz and quantized using 16 bits/sample) at bit rates of 64-78 Kbits/sec. Our implementation uses two ASPI Elf boards, each of which is built around a TI TMS230C31 DSP chip. The time required for encoding of a mono CD signal is about 92 percent of real time and that for decoding about 61 percent.
Application of Computer Aided Mathematics Teaching in a Secondary School
ERIC Educational Resources Information Center
Yenitepe, Mehmet Emin; Karadag, Zekeriya
2003-01-01
This is a case study that examines the effect of using presentations developed by teacher in addition to using commercially produced educational software CD-ROM in Audio-Visual Room/Computer Laboratory after classroom teaching, on students' academic achievement, as a method of Teaching Mathematics compared with only classroom teaching or after…
Oral Computer-Mediated Interaction between L2 Learners: It's about Time!
ERIC Educational Resources Information Center
Yanguas, Inigo
2010-01-01
This study explores task-based, synchronous oral computer-mediated communication (CMC) among intermediate-level learners of Spanish. In particular, this paper examines (a) how learners in video and audio CMC groups negotiate for meaning during task-based interaction, (b) possible differences between both oral CMC modes and traditional face-to-face…
Edwards, Sandra L; Slattery, Martha L; Murtaugh, Maureen A; Edwards, Roger L; Bryner, James; Pearson, Mindy; Rogers, Amy; Edwards, Alison M; Tom-Orme, Lillian
2007-06-01
This article describes the development and usability of an audio computer-assisted self-interviewing (ACASI) questionnaire created to collect dietary, physical activity, medical history, and other lifestyle data in a population of American Indians. Study participants were part of a cohort of American Indians living in the southwestern United States. Data were collected between March 2004 and July 2005. Information for evaluating questionnaire usability and acceptability was collected from three different sources: baseline study data, auxiliary background data, and a short questionnaire administered to a subset of study participants. For the subset of participants, 39.6% reported not having used a computer in the past year. The ACASI questionnaires were well accepted: 96.0% of the subset of participants reported finding them enjoyable to use, 97.2% reported that they were easy to use, and 82.6% preferred them for future questionnaires. A lower educational level and infrequent computer use in the past year were predictors of having usability trouble. These results indicate that the ACASI questionnaire is both an acceptable and a preferable mode of data collection in this population.
Optimal Window and Lattice in Gabor Transform. Application to Audio Analysis.
Lachambre, Helene; Ricaud, Benjamin; Stempfel, Guillaume; Torrésani, Bruno; Wiesmeyr, Christoph; Onchis-Moaca, Darian
2015-01-01
This article deals with the use of optimal lattice and optimal window in Discrete Gabor Transform computation. In the case of a generalized Gaussian window, extending earlier contributions, we introduce an additional local window adaptation technique for non-stationary signals. We illustrate our approach and the earlier one by addressing three time-frequency analysis problems to show the improvements achieved by the use of optimal lattice and window: close frequencies distinction, frequency estimation and SNR estimation. The results are presented, when possible, with real world audio signals.
Mosier, Jarrod; Joseph, Bellal; Sakles, John C
2013-02-01
Since the first remote intubation with telemedicine guidance, wireless technology has advanced to enable more portable methods of telemedicine involvement in remote airway management. Three voice over Internet protocol (VoIP) services were evaluated for quality of image transmitted, data lag, and audio quality with remotely observed and assisted intubations in an academic emergency department. The VoIP clients evaluated were Apple (Cupertino, CA) FaceTime(®), Skype™ (a division of Microsoft, Luxembourg City, Luxembourg), and Tango(®) (TangoMe, Palo Alto, CA). Each client was tested over a Wi-Fi network as well as cellular third generation (3G) (Skype and Tango). All three VoIP clients provided acceptable image and audio quality. There is a significant data lag in image transmission and quality when VoIP clients are used over cellular broadband (3G) compared with Wi-Fi. Portable remote telemedicine guidance is possible with newer technology devices such as a smartphone or tablet, as well as VoIP clients used over Wi-Fi or cellular broadband.
Quality Control for Interviews to Obtain Dietary Recalls from Children for Research Studies
SHAFFER, NICOLE M.; THOMPSON, WILLIAM O.; BAGLIO, MICHELLE L.; GUINN, CAROLINE H.; FRYE, FRANCESCA H. A.
2005-01-01
Quality control is an important aspect of a study because the quality of data collected provides a foundation for the conclusions drawn from the study. For studies that include interviews, establishing quality control for interviews is critical in ascertaining whether interviews are conducted according to protocol. Despite the importance of quality control for interviews, few studies adequately document the quality control procedures used during data collection. This article reviews quality control for interviews and describes methods and results of quality control for interviews from two of our studies regarding the accuracy of children's dietary recalls; the focus is on quality control regarding interviewer performance during the interview, and examples are provided from studies with children. For our two studies, every interview was audio recorded and transcribed. The audio recording and typed transcript from one interview conducted by each research dietitian either weekly or daily were randomly selected and reviewed by another research dietitian, who completed a standardized quality control for interviews checklist. Major strengths of the methods of quality control for interviews in our two studies include: (a) interviews obtained for data collection were randomly selected for quality control for interviews, and (b) quality control for interviews was assessed on a regular basis throughout data collection. The methods of quality control for interviews described may help researchers design appropriate methods of quality control for interviews for future studies. PMID:15389417
Subjective Audio Quality over a Secure IEEE 802.11n Draft 2.0 Wireless Local Area Network
2009-03-01
hereafter referred to as 802.11) provide users with mobile connectivity without the need for expensive and inflexible wiring. The 802.11n extension, for...through another protocol, such as Secure / Multipurpose Internet Mail Extensions ( S /MIME). SDPS is, therefore, not a complete solution for secure key...number of packets per second (“Pkts/ s ”) are visible. Audio recordings are taken at AFIT within range of several other 802.11g APs as shown in Figure
77 FR 74543 - Federal Advisory Committee Meeting
Federal Register 2010, 2011, 2012, 2013, 2014
2012-12-14
.../411593 . Dial-in: After you've connected your computer, audio connection instructions will be presented... the status of current research projects. FOR FURTHER INFORMATION CONTACT: The meeting is open to the...
Using Speech Recognition to Enhance the Tongue Drive System Functionality in Computer Access
Huo, Xueliang; Ghovanloo, Maysam
2013-01-01
Tongue Drive System (TDS) is a wireless tongue operated assistive technology (AT), which can enable people with severe physical disabilities to access computers and drive powered wheelchairs using their volitional tongue movements. TDS offers six discrete commands, simultaneously available to the users, for pointing and typing as a substitute for mouse and keyboard in computer access, respectively. To enhance the TDS performance in typing, we have added a microphone, an audio codec, and a wireless audio link to its readily available 3-axial magnetic sensor array, and combined it with a commercially available speech recognition software, the Dragon Naturally Speaking, which is regarded as one of the most efficient ways for text entry. Our preliminary evaluations indicate that the combined TDS and speech recognition technologies can provide end users with significantly higher performance than using each technology alone, particularly in completing tasks that require both pointing and text entry, such as web surfing. PMID:22255801
NASA Astrophysics Data System (ADS)
Pallone, Arthur
Necessity often leads to inspiration. Such was the case when a traditional amplifier quit working during the collection of an alpha particle spectrum. I had a 15 battery-powered audio amplifier in my box of project electronics so I connected it between the preamplifier and the multichannel analyzer. The alpha particle spectrum that appeared on the computer screen matched expectations even without correcting for impedance mismatches. Encouraged by this outcome, I have begun to systematically replace each of the parts in a traditional charged particle spectrometer with audio and video components available through consumer electronics stores with the goal of producing an inexpensive charged particle spectrometer for use in education and research. Hopefully my successes, setbacks, and results to date described in this presentation will inform and inspire others.
New Literacy Tools for Adults.
ERIC Educational Resources Information Center
Anderson, Jonathan
1990-01-01
Describes an Australian national study of technologies used for adult literacy: traditional technologies (print, radio, television, audio and videotape, teleconferencing, and computers) and new generation technologies (laser discs, CD-ROM, videodiscs, and hypermedia). (SK)
Satellite sound broadcasting system, portable reception
NASA Technical Reports Server (NTRS)
Golshan, Nasser; Vaisnys, Arvydas
1990-01-01
Studies are underway at JPL in the emerging area of Satellite Sound Broadcast Service (SSBS) for direct reception by low cost portable, semi portable, mobile and fixed radio receivers. This paper addresses the portable reception of digital broadcasting of monophonic audio with source material band limited to 5 KHz (source audio comparable to commercial AM broadcasting). The proposed system provides transmission robustness, uniformity of performance over the coverage area and excellent frequency reuse. Propagation problems associated with indoor portable reception are considered in detail and innovative antenna concepts are suggested to mitigate these problems. It is shown that, with the marriage of proper technologies a single medium power satellite can provide substantial direct satellite audio broadcast capability to CONUS in UHF or L Bands, for high quality portable indoor reception by low cost radio receivers.
Microcomputer Software Development: New Strategies for a New Technology.
ERIC Educational Resources Information Center
Kehrberg, Kent T.
1979-01-01
Provides a guide for the development of educational computer programs for use on microcomputers. Making use of the features of microcomputers, including visual, audio, and tactile techniques, is encouraged. (Author/IRT)
Task-Based Oral Computer-Mediated Communication and L2 Vocabulary Acquisition
ERIC Educational Resources Information Center
Yanguas, Inigo
2012-01-01
The present study adds to the computer-mediated communication (CMC) literature by exploring oral learner-to-learner interaction using Skype, a free and widely used Internet software program. In particular, this task-based study has a two-fold goal. Firstly, it explores possible differences between two modes of oral CMC (audio and video) and…
That's Infotainment!: How to Create Your Own Screencasts
ERIC Educational Resources Information Center
Kroski, Ellyssa
2009-01-01
Screencasts are videos that record the actions that take place on the computer screen, most often including a narrative audio track, in order to demonstrate various computer-related tasks, such as how to use a software program or navigate a certain Web site. All that is needed is a standard microphone and screen recording software, which can be…
Power Monitoring Using the Raspberry Pi
ERIC Educational Resources Information Center
Snyder, Robin M.
2014-01-01
The Raspberry Pi is a credit card size low powered compute board with Ethernet connection, HDMI video output, audio, full Linux operating system run from an SD card, and more, all for $45. With cables, SD card, etc., the cost is about $70. Originally designed to help teach computer science principles to low income children and students, the Pi has…
ERIC Educational Resources Information Center
vonFeldt, James R.
The development of a prototype system is described which merges the strengths of computer assisted instruction, data gathering, interactive learning, individualized instruction, and the motion in color, and audio features of television. Creation of the prototype system will allow testing of both TV and interactive CAI/TV strategies in auditory and…
ERIC Educational Resources Information Center
Hostetler, Jerry C.; Englert, Duwayne C.
1987-01-01
Presents description of an interface device which ties in microcomputers and slide/tape presentations for computer assisted instruction. Highlights include the use of this technology in an introductory undergraduate zoology course; a discussion of authoring languages with emphasis on SuperPILOT; and hardware and software design for the interface.…
Recording high quality speech during tagged cine-MRI studies using a fiber optic microphone.
NessAiver, Moriel S; Stone, Maureen; Parthasarathy, Vijay; Kahana, Yuvi; Paritsky, Alexander; Paritsky, Alex
2006-01-01
To investigate the feasibility of obtaining high quality speech recordings during cine imaging of tongue movement using a fiber optic microphone. A Complementary Spatial Modulation of Magnetization (C-SPAMM) tagged cine sequence triggered by an electrocardiogram (ECG) simulator was used to image a volunteer while speaking the syllable pairs /a/-/u/, /i/-/u/, and the words "golly" and "Tamil" in sync with the imaging sequence. A noise-canceling, optical microphone was fastened approximately 1-2 inches above the mouth of the volunteer. The microphone was attached via optical fiber to a laptop computer, where the speech was sampled at 44.1 kHz. A reference recording of gradient activity with no speech was subtracted from target recordings. Good quality speech was discernible above the background gradient sound using the fiber optic microphone without reference subtraction. The audio waveform of gradient activity was extremely stable and reproducible. Subtraction of the reference gradient recording further reduced gradient noise by roughly 21 dB, resulting in exceptionally high quality speech waveforms. It is possible to obtain high quality speech recordings using an optical microphone even during exceptionally loud cine imaging sequences. This opens up the possibility of more elaborate MRI studies of speech including spectral analysis of the speech signal in all types of MRI.
NASA Astrophysics Data System (ADS)
Ismail M., E.; Mahazir I., Irwan; Othman, H.; Amiruddin M., H.; Ariffin, A.
2017-05-01
The rapid development of information technology today has given a new breath toward usage of computer in education. One of the increasingly popular nowadays is a multimedia technology that merges a variety of media such as text, graphics, animation, video and audio controlled by a computer. With this technology, a wide range of multimedia element can be developed to improve the quality of education. For that reason, this study aims to investigate the use of multimedia element based on animated video that was developed for Engineering Drawing subject according to the syllabus of Vocational College of Malaysia. The design for this study was a survey method using a quantitative approach and involved 30 respondents from Industrial Machining students. The instruments used in study is questionnaire with correlation coefficient value (0.83), calculated on Alpha-Cronbach. Data was collected and analyzed descriptive analyzed using SPSS. The study found that multimedia element for animation video was use significant have capable to increase imagination and visualization of student. The implications of this study provide information of use of multimedia element will student effect imagination and visualization. In general, these findings contribute to the formation of multimedia element of materials appropriate to enhance the quality of learning material for engineering drawing.
Architectures for single-chip image computing
NASA Astrophysics Data System (ADS)
Gove, Robert J.
1992-04-01
This paper will focus on the architectures of VLSI programmable processing components for image computing applications. TI, the maker of industry-leading RISC, DSP, and graphics components, has developed an architecture for a new-generation of image processors capable of implementing a plurality of image, graphics, video, and audio computing functions. We will show that the use of a single-chip heterogeneous MIMD parallel architecture best suits this class of processors--those which will dominate the desktop multimedia, document imaging, computer graphics, and visualization systems of this decade.
Free oscilloscope web app using a computer mic, built-in sound library, or your own files
NASA Astrophysics Data System (ADS)
Ball, Edward; Ruiz, Frances; Ruiz, Michael J.
2017-07-01
We have developed an online oscilloscope program which allows users to see waveforms by utilizing their computer microphones, selecting from our library of over 30 audio files, and opening any *.mp3 or *.wav file on their computers. The oscilloscope displays real-time signals against time. The oscilloscope has been calibrated so one can make accurate frequency measurements of periodic waves to within 1%. The web app is ideal for computer projection in class.
Bai, Mingsian R; Wen, Jheng-Ciang; Hsu, Hoshen; Hua, Yi-Hsin; Hsieh, Yu-Hao
2014-10-01
A sound reconstruction system is proposed for audio reproduction with extended sweet spot and reduced reflections. An equivalent source method (ESM)-based sound field synthesis (SFS) approach, with the aid of dark zone minimization is adopted in the study. Conventional SFS that is based on the free-field assumption suffers from synthesis error due to boundary reflections. To tackle the problem, the proposed system utilizes convex optimization in designing array filters with both reproduction performance and acoustic contrast taken into consideration. Control points are deployed in the dark zone to minimize the reflections from the walls. Two approaches are employed to constrain the pressure and velocity in the dark zone. Pressure matching error (PME) and acoustic contrast (AC) are used as performance measures in simulations and experiments for a rectangular loudspeaker array. Perceptual Evaluation of Audio Quality (PEAQ) is also used to assess the audio reproduction quality. The results show that the pressure-constrained (PC) method yields better acoustic contrast, but poorer reproduction performance than the pressure-velocity constrained (PVC) method. A subjective listening test also indicates that the PVC method is the preferred method in a live room.
Efficient Geometric Sound Propagation Using Visibility Culling
NASA Astrophysics Data System (ADS)
Chandak, Anish
2011-07-01
Simulating propagation of sound can improve the sense of realism in interactive applications such as video games and can lead to better designs in engineering applications such as architectural acoustics. In this thesis, we present geometric sound propagation techniques which are faster than prior methods and map well to upcoming parallel multi-core CPUs. We model specular reflections by using the image-source method and model finite-edge diffraction by using the well-known Biot-Tolstoy-Medwin (BTM) model. We accelerate the computation of specular reflections by applying novel visibility algorithms, FastV and AD-Frustum, which compute visibility from a point. We accelerate finite-edge diffraction modeling by applying a novel visibility algorithm which computes visibility from a region. Our visibility algorithms are based on frustum tracing and exploit recent advances in fast ray-hierarchy intersections, data-parallel computations, and scalable, multi-core algorithms. The AD-Frustum algorithm adapts its computation to the scene complexity and allows small errors in computing specular reflection paths for higher computational efficiency. FastV and our visibility algorithm from a region are general, object-space, conservative visibility algorithms that together significantly reduce the number of image sources compared to other techniques while preserving the same accuracy. Our geometric propagation algorithms are an order of magnitude faster than prior approaches for modeling specular reflections and two to ten times faster for modeling finite-edge diffraction. Our algorithms are interactive, scale almost linearly on multi-core CPUs, and can handle large, complex, and dynamic scenes. We also compare the accuracy of our sound propagation algorithms with other methods. Once sound propagation is performed, it is desirable to listen to the propagated sound in interactive and engineering applications. We can generate smooth, artifact-free output audio signals by applying efficient audio-processing algorithms. We also present the first efficient audio-processing algorithm for scenarios with simultaneously moving source and moving receiver (MS-MR) which incurs less than 25% overhead compared to static source and moving receiver (SS-MR) or moving source and static receiver (MS-SR) scenario.
Audio visual speech source separation via improved context dependent association model
NASA Astrophysics Data System (ADS)
Kazemi, Alireza; Boostani, Reza; Sobhanmanesh, Fariborz
2014-12-01
In this paper, we exploit the non-linear relation between a speech source and its associated lip video as a source of extra information to propose an improved audio-visual speech source separation (AVSS) algorithm. The audio-visual association is modeled using a neural associator which estimates the visual lip parameters from a temporal context of acoustic observation frames. We define an objective function based on mean square error (MSE) measure between estimated and target visual parameters. This function is minimized for estimation of the de-mixing vector/filters to separate the relevant source from linear instantaneous or time-domain convolutive mixtures. We have also proposed a hybrid criterion which uses AV coherency together with kurtosis as a non-Gaussianity measure. Experimental results are presented and compared in terms of visually relevant speech detection accuracy and output signal-to-interference ratio (SIR) of source separation. The suggested audio-visual model significantly improves relevant speech classification accuracy compared to existing GMM-based model and the proposed AVSS algorithm improves the speech separation quality compared to reference ICA- and AVSS-based methods.
Stropahl, Maren; Schellhardt, Sebastian; Debener, Stefan
2017-06-01
The concurrent presentation of different auditory and visual syllables may result in the perception of a third syllable, reflecting an illusory fusion of visual and auditory information. This well-known McGurk effect is frequently used for the study of audio-visual integration. Recently, it was shown that the McGurk effect is strongly stimulus-dependent, which complicates comparisons across perceivers and inferences across studies. To overcome this limitation, we developed the freely available Oldenburg audio-visual speech stimuli (OLAVS), consisting of 8 different talkers and 12 different syllable combinations. The quality of the OLAVS set was evaluated with 24 normal-hearing subjects. All 96 stimuli were characterized based on their stimulus disparity, which was obtained from a probabilistic model (cf. Magnotti & Beauchamp, 2015). Moreover, the McGurk effect was studied in eight adult cochlear implant (CI) users. By applying the individual, stimulus-independent parameters of the probabilistic model, the predicted effect of stronger audio-visual integration in CI users could be confirmed, demonstrating the validity of the new stimulus material.
Perceptually controlled doping for audio source separation
NASA Astrophysics Data System (ADS)
Mahé, Gaël; Nadalin, Everton Z.; Suyama, Ricardo; Romano, João MT
2014-12-01
The separation of an underdetermined audio mixture can be performed through sparse component analysis (SCA) that relies however on the strong hypothesis that source signals are sparse in some domain. To overcome this difficulty in the case where the original sources are available before the mixing process, the informed source separation (ISS) embeds in the mixture a watermark, which information can help a further separation. Though powerful, this technique is generally specific to a particular mixing setup and may be compromised by an additional bitrate compression stage. Thus, instead of watermarking, we propose a `doping' method that makes the time-frequency representation of each source more sparse, while preserving its audio quality. This method is based on an iterative decrease of the distance between the distribution of the signal and a target sparse distribution, under a perceptual constraint. We aim to show that the proposed approach is robust to audio coding and that the use of the sparsified signals improves the source separation, in comparison with the original sources. In this work, the analysis is made only in instantaneous mixtures and focused on voice sources.
Comparison of audio and audiovisual measures of adult stuttering: Implications for clinical trials.
O'Brian, Sue; Jones, Mark; Onslow, Mark; Packman, Ann; Menzies, Ross; Lowe, Robyn
2015-04-15
This study investigated whether measures of percentage syllables stuttered (%SS) and stuttering severity ratings with a 9-point scale differ when made from audiovisual compared with audio-only recordings. Four experienced speech-language pathologists measured %SS and assigned stuttering severity ratings to 10-minute audiovisual and audio-only recordings of 36 adults. There was a mean 18% increase in %SS scores when samples were presented in audiovisual compared with audio-only mode. This result was consistent across both higher and lower %SS scores and was found to be directly attributable to counts of stuttered syllables rather than the total number of syllables. There was no significant difference between stuttering severity ratings made from the two modes. In clinical trials research, when using %SS as the primary outcome measure, audiovisual samples would be preferred as long as clear, good quality, front-on images can be easily captured. Alternatively, stuttering severity ratings may be a more valid measure to use as they correlate well with %SS and values are not influenced by the presentation mode.
Coexistence issues for a 2.4 GHz wireless audio streaming in presence of bluetooth paging and WLAN
NASA Astrophysics Data System (ADS)
Pfeiffer, F.; Rashwan, M.; Biebl, E.; Napholz, B.
2015-11-01
Nowadays, customers expect to integrate their mobile electronic devices (smartphones and laptops) in a vehicle to form a wireless network. Typically, IEEE 802.11 is used to provide a high-speed wireless local area network (WLAN) and Bluetooth is used for cable replacement applications in a wireless personal area network (PAN). In addition, Daimler uses KLEER as third wireless technology in the unlicensed (UL) 2.4 GHz-ISM-band to transmit full CD-quality digital audio. As Bluetooth, IEEE 802.11 and KLEER are operating in the same frequency band, it has to be ensured that all three technologies can be used simultaneously without interference. In this paper, we focus on the impact of Bluetooth and IEEE 802.11 as interferer in presence of a KLEER audio transmission.
Trelease, R B
1996-01-01
Advances in computer visualization and user interface technologies have enabled development of "virtual reality" programs that allow users to perceive and to interact with objects in artificial three-dimensional environments. Such technologies were used to create an image database and program for studying the human skull, a specimen that has become increasingly expensive and scarce. Stereoscopic image pairs of a museum-quality skull were digitized from multiple views. For each view, the stereo pairs were interlaced into a single, field-sequential stereoscopic picture using an image processing program. The resulting interlaced image files are organized in an interactive multimedia program. At run-time, gray-scale 3-D images are displayed on a large-screen computer monitor and observed through liquid-crystal shutter goggles. Users can then control the program and change views with a mouse and cursor to point-and-click on screen-level control words ("buttons"). For each view of the skull, an ID control button can be used to overlay pointers and captions for important structures. Pointing and clicking on "hidden buttons" overlying certain structures triggers digitized audio spoken word descriptions or mini lectures.
NASA Astrophysics Data System (ADS)
O'Donnell, Michael J.; Bisnovatyi, Ilia
2000-11-01
Computing practice today depends on visual output to drive almost all user interaction. Other senses, such as audition, may be totally neglected, or used tangentially, or used in highly restricted specialized ways. We have excellent audio rendering through D-A conversion, but we lack rich general facilities for modeling and manipulating sound comparable in quality and flexibility to graphics. We need coordinated research in several disciplines to improve the use of sound as an interactive information channel. Incremental and separate improvements in synthesis, analysis, speech processing, audiology, acoustics, music, etc. will not alone produce the radical progress that we seek in sonic practice. We also need to create a new central topic of study in digital audio research. The new topic will assimilate the contributions of different disciplines on a common foundation. The key central concept that we lack is sound as a general-purpose information channel. We must investigate the structure of this information channel, which is driven by the cooperative development of auditory perception and physical sound production. Particular audible encodings, such as speech and music, illuminate sonic information by example, but they are no more sufficient for a characterization than typography is sufficient for characterization of visual information. To develop this new conceptual topic of sonic information structure, we need to integrate insights from a number of different disciplines that deal with sound. In particular, we need to coordinate central and foundational studies of the representational models of sound with specific applications that illuminate the good and bad qualities of these models. Each natural or artificial process that generates informative sound, and each perceptual mechanism that derives information from sound, will teach us something about the right structure to attribute to the sound itself. The new Sound topic will combine the work of computer scientists with that of numerical mathematicians studying sonification, psychologists, linguists, bioacousticians, and musicians to illuminate the structure of sound from different angles. Each of these disciplines deals with the use of sound to carry a different sort of information, under different requirements and constraints. By combining their insights, we can learn to understand of the structure of sound in general.
Direct Measurement of the Speed of Sound Using a Microphone and a Speaker
ERIC Educational Resources Information Center
Gómez-Tejedor, José A.; Castro-Palacio, Juan C.; Monsoriu, Juan A.
2014-01-01
We present a simple and accurate experiment to obtain the speed of sound in air using a conventional speaker and a microphone connected to a computer. A free open source digital audio editor and recording computer software application allows determination of the time-of-flight of the wave for different distances, from which the speed of sound is…
Multimedia Instruction Puts Teachers in the Director's Chair.
ERIC Educational Resources Information Center
Trotter, Andrew
1990-01-01
Teachers can produce and direct their own instructional videos using computer-driven multimedia. Outlines the basics in combining audio and video technologies to produce videotapes that mix animated and still graphics, sound, and full-motion video. (MLF)
Message Modality and Source Credibility Can Interact to Affect Argument Processing.
ERIC Educational Resources Information Center
Booth-Butterfield, Steve; Gutowski, Christine
1993-01-01
Extends previous modality and source cue studies by manipulating argument quality. Randomly assigned college students by class to an argument quality by source attribute by modality factorial experiment. Finds the print mode produces only argument main effects, and audio and video modes produce argument by cue interactions. Finds data inconsistent…
Instructional Activities and the Quality of Language in Chilean Preschool Classrooms
ERIC Educational Resources Information Center
Strasser, Katherine; Darricades, Michelle; Mendive, Susana; Barra, Gabriela
2018-01-01
Research Findings: This study examines the association between preschool classroom activity and the quality of the language spoken by teachers and children. Eighteen classrooms serving low-income children between the ages of 3 and 4 in Santiago de Chile were audio-recorded during one morning shift. Recordings were transcribed and segmented into…
The VTLA System of Course Delivery and Faculty Development in Materials Education
NASA Technical Reports Server (NTRS)
Berrettini, Robert; Roy, Rustum
1996-01-01
There is a national need for high-quality, upper division courses that address critical topics in materials synthesis, particularly those beyond the present expertise of the typical university department's faculty. A new project has been started to test a novel distance education and faculty development system, called Video Tape Live Audio (VTLA). This, if successful, would at once enlarge the national Materials Science and Engineering (MSE) student cohort studying material synthesis and develop faculty expertise at the receiving sites. The mechanics for the VTLA scheme are as follows: A course is designed in the field selected for emphasis and for which there is likely to be considerable demand, in this example 'Ceramic Materials Synthesis: Theory and Case Studies'. One of the very best researcher/teachers records lectures of TV studio quality with appropriate visuals. Universities and colleges which wish to offer the course agree to offer it at the same hour at least once a week. The videotaped lectures and accompanying text, readings and visuals are shipped to the professor in charge, who has an appropriate background. The professor arranges the classroom TV presentation equipment and supervises the course. Video lectures are played during regular course hours twice a week with time for discussion by the supervising professor. Typically the third weekly classroom period is scheduled by all sites at a common designated hour, during which the course author/presenter answers questions, provides greater depth, etc. on a live audio link to all course sites. Questions are submitted by fax and e-mail prior to the audio tutorial. coordinating professors at various sites have separate audio teleconferences at the beginning and end of the course, dealing with the philosophical and pedagogical approach to the course, content and mechanics. Following service once or twice as an 'apprentice' to the course, the coordinating professors may then offer it without the necessity of the live audio tutorial.
Automatic summarization of soccer highlights using audio-visual descriptors.
Raventós, A; Quijada, R; Torres, Luis; Tarrés, Francesc
2015-01-01
Automatic summarization generation of sports video content has been object of great interest for many years. Although semantic descriptions techniques have been proposed, many of the approaches still rely on low-level video descriptors that render quite limited results due to the complexity of the problem and to the low capability of the descriptors to represent semantic content. In this paper, a new approach for automatic highlights summarization generation of soccer videos using audio-visual descriptors is presented. The approach is based on the segmentation of the video sequence into shots that will be further analyzed to determine its relevance and interest. Of special interest in the approach is the use of the audio information that provides additional robustness to the overall performance of the summarization system. For every video shot a set of low and mid level audio-visual descriptors are computed and lately adequately combined in order to obtain different relevance measures based on empirical knowledge rules. The final summary is generated by selecting those shots with highest interest according to the specifications of the user and the results of relevance measures. A variety of results are presented with real soccer video sequences that prove the validity of the approach.
de la Torre, A N; Castaneda, I; Ahmad, M; Ekholy, N; Tham, N; Herrera, I B; Beaty, P; Malapero, R J; Ayoub, F; Slim, J; Johnson, M B
2017-12-01
Intravenous drug use and sexual practices account for 60% of hepatitis C (HCV) and B (HBV) infection. Disclosing these activities can be embarrassing and reduce risk reporting, blood testing and diagnosis. In diagnosed patients, linkage to care remains a challenge. Audio-computer-assisted survey interview (Audio-CASI) was used to guide HCV and HBV infection testing in urban clinics. Risk reporting, blood testing and serology results were compared to historical controls. A patient navigator (PN) followed up blood test results and provided patients with positive serology linkage to care (LTC). Of 1932 patients surveyed, 574 (30%) were at risk for chronic viral hepatitis. A total of 254 (44.3%) patients were tested, 34 (13.5%) had serology warranting treatment evaluation, and 64% required HBV vaccination. Of 16 patients with infection, seven HCV and three HBV patients started treatment following patient LTC. Of 146 HBV-naïve patients, 70 completed vaccination. About 75% and 49% of HCV antibody and HBV surface antigen-positive patients were born between 1945 and 1965. Subsequently, automated HCV testing of patients born between 1945 and 1965 was built into our hospital electronic medical records. Average monthly HCV antibody testing increased from 245 (January-June) to 1187 (July-October). Patient navigator directed LTC for HCV antibody-positive patients was 61.6%. In conclusion, audio-CASI can identify patients at risk for HCV or HBV infection and those in need of HBV vaccination in urban medical clinics. Although blood testing once a patient is identified at risk for infection needs to increase, a PN is useful to provide LTC of newly diagnosed patients. © 2017 John Wiley & Sons Ltd.
ERIC Educational Resources Information Center
Falk, Howard
1998-01-01
Discussion of CD (compact disc) recorders describes recording applications, including storing large graphic files, creating audio CDs, and storing material downloaded from the Internet; backing up files; lifespan; CD recording formats; continuous recording; recording software; recorder media; vulnerability of CDs; basic computer requirements; and…
Validation of air traffic controller workload models
DOT National Transportation Integrated Search
1979-09-01
During the past several years, computer models have been developed for off-site : estimat ion of control ler's workload. The inputs to these models are audio and : digital data normally recorded at an Air Route Traffic Control Center (ARTCC). : This ...
Gedanken Experiments in Educational Cost Effectiveness
ERIC Educational Resources Information Center
Brudner, Harvey J.
1978-01-01
Discusses the effectiveness of cost determining techniques in education. The areas discussed are: education and management; cost-effectiveness models; figures of merit determination; and the implications as they relate to the areas of audio-visual and computer educational technology. (Author/GA)
78 FR 29063 - Survey of Urban Rates for Fixed Voice and Fixed Broadband Residential Services
Federal Register 2010, 2011, 2012, 2013, 2014
2013-05-17
... in alternative formats (computer diskette, large print, audio record, and Braille). Persons with... Company Name: Provider FRN (used on MONTH DAY, YEAR Form 477): Provider Study Area Code (if current USF...
Measuring the rebound resilience of a bouncing ball
NASA Astrophysics Data System (ADS)
Wadhwa, Ajay
2012-09-01
Some balls which are made of high-quality rubber (an elastomeric) material, such as tennis or squash balls, could be used for the determination of an important property of such materials called resilience. Since a bouncing ball involves a single impact we call this property 'rebound resilience' and express it as the ratio of the rebound height to the initial drop height of the ball. We determine the rebound resilience for three different types of ball by calculating the coefficient of restitution of the ball-surface combination from the experimentally measurable physical quantities, such as initial drop height and time interval between successive bounces. Using these we also determine the contact time of balls with the surface of impact. For measurements we have used audio, motion and surface-temperature sensors that were interfaced through a USB port with a computer.
Efficient audio signal processing for embedded systems
NASA Astrophysics Data System (ADS)
Chiu, Leung Kin
As mobile platforms continue to pack on more computational power, electronics manufacturers start to differentiate their products by enhancing the audio features. However, consumers also demand smaller devices that could operate for longer time, hence imposing design constraints. In this research, we investigate two design strategies that would allow us to efficiently process audio signals on embedded systems such as mobile phones and portable electronics. In the first strategy, we exploit properties of the human auditory system to process audio signals. We designed a sound enhancement algorithm to make piezoelectric loudspeakers sound ”richer" and "fuller." Piezoelectric speakers have a small form factor but exhibit poor response in the low-frequency region. In the algorithm, we combine psychoacoustic bass extension and dynamic range compression to improve the perceived bass coming out from the tiny speakers. We also developed an audio energy reduction algorithm for loudspeaker power management. The perceptually transparent algorithm extends the battery life of mobile devices and prevents thermal damage in speakers. This method is similar to audio compression algorithms, which encode audio signals in such a ways that the compression artifacts are not easily perceivable. Instead of reducing the storage space, however, we suppress the audio contents that are below the hearing threshold, therefore reducing the signal energy. In the second strategy, we use low-power analog circuits to process the signal before digitizing it. We designed an analog front-end for sound detection and implemented it on a field programmable analog array (FPAA). The system is an example of an analog-to-information converter. The sound classifier front-end can be used in a wide range of applications because programmable floating-gate transistors are employed to store classifier weights. Moreover, we incorporated a feature selection algorithm to simplify the analog front-end. A machine learning algorithm AdaBoost is used to select the most relevant features for a particular sound detection application. In this classifier architecture, we combine simple "base" analog classifiers to form a strong one. We also designed the circuits to implement the AdaBoost-based analog classifier.
Impact of Audio-Coaching on the Position of Lung Tumors
DOE Office of Scientific and Technical Information (OSTI.GOV)
Haasbeek, Cornelis J.A.; Spoelstra, Femke; Lagerwaard, Frank J.
2008-07-15
Purpose: Respiration-induced organ motion is a major source of positional, or geometric, uncertainty in thoracic radiotherapy. Interventions to mitigate the impact of motion include audio-coached respiration-gated radiotherapy (RGRT). To assess the impact of coaching on average tumor position during gating, we analyzed four-dimensional computed tomography (4DCT) scans performed both with and without audio-coaching. Methods and Materials: Our RGRT protocol requires that an audio-coached 4DCT scan is performed when the initial free-breathing 4DCT indicates a potential benefit with gating. We retrospectively analyzed 22 such paired scans in patients with well-circumscribed tumors. Changes in lung volume and position of internal target volumesmore » (ITV) generated in three consecutive respiratory phases at both end-inspiration and end-expiration were analyzed. Results: Audio-coaching increased end-inspiration lung volumes by a mean of 10.2% (range, -13% to +43%) when compared with free breathing (p = 0.001). The mean three-dimensional displacement of the center of ITV was 3.6 mm (SD, 2.5; range, 0.3-9.6mm), mainly caused by displacement in the craniocaudal direction. Displacement of ITV caused by coaching was more than 5 mm in 5 patients, all of whom were in the subgroup of 9 patients showing total tumor motion of 10 mm or more during both coached and uncoached breathing. Comparable ITV displacements were observed at end-expiration phases of the 4DCT. Conclusions: Differences in ITV position exceeding 5 mm between coached and uncoached 4DCT scans were detected in up to 56% of mobile tumors. Both end-inspiration and end-expiration RGRT were susceptible to displacements. This indicates that the method of audio-coaching should remain unchanged throughout the course of treatment.« less
Blogging Is So Last Year--Now Podcasting Is Hot. Online Treasures
ERIC Educational Resources Information Center
Balas, Janet L.
2005-01-01
This column describes the latest innovation in iPod use, podcasts, which are audio files that can be downloaded and played either through a computer or an MP3 player such as an iPod. Because they are feeds, users can subscribe to podcasts and have them downloaded automatically to their computers. The author discusses ways in which podcasts may be…
Computer-Mediated Communication: Decisionmaking and Informal Interaction.
1985-02-15
the context of a project called the Computer-Administered Panel Study L (CAPS). Respondents in the project were 96 UNC undergraduate students . The...project consisted of the administration of over 100 questionnaires and experimental units to this panel of undergraduates who reported to weekly sessions...humanely (eg., are more likely to ignore or insult) via the more distant media. Audio-only communications are more likely to be depersonalized
ERIC Educational Resources Information Center
Library of Congress, Washington, DC. Congressional Research Service.
This summary of the combined Hearing and Workshop on Applications of Computer-Based Information Systems and Services in Agriculture (May 19-20, 1982) offers an overview of the ways in which information technology--computers, telecommunications, microforms, word processing, video and audio devices--may be utilized by American farmers and ranchers.…
Digital video technology, today and tomorrow
NASA Astrophysics Data System (ADS)
Liberman, J.
1994-10-01
Digital video is probably computing's fastest moving technology today. Just three years ago, the zenith of digital video technology on the PC was the successful marriage of digital text and graphics with analog audio and video by means of expensive analog laser disc players and video overlay boards. The state of the art involves two different approaches to fully digital video on computers: hardware-assisted and software-only solutions.
A Robust Zero-Watermarking Algorithm for Audio
NASA Astrophysics Data System (ADS)
Chen, Ning; Zhu, Jie
2007-12-01
In traditional watermarking algorithms, the insertion of watermark into the host signal inevitably introduces some perceptible quality degradation. Another problem is the inherent conflict between imperceptibility and robustness. Zero-watermarking technique can solve these problems successfully. Instead of embedding watermark, the zero-watermarking technique extracts some essential characteristics from the host signal and uses them for watermark detection. However, most of the available zero-watermarking schemes are designed for still image and their robustness is not satisfactory. In this paper, an efficient and robust zero-watermarking technique for audio signal is presented. The multiresolution characteristic of discrete wavelet transform (DWT), the energy compression characteristic of discrete cosine transform (DCT), and the Gaussian noise suppression property of higher-order cumulant are combined to extract essential features from the host audio signal and they are then used for watermark recovery. Simulation results demonstrate the effectiveness of our scheme in terms of inaudibility, detection reliability, and robustness.
NASA Astrophysics Data System (ADS)
Aishwariya, A.; Pallavi Sudhir, Gulavani; Garg, Nemesa; Karthikeyan, B.
2017-11-01
A body worn camera is small video camera worn on the body, typically used by police officers to record arrests, evidence from crime scenes. It helps preventing and resolving complaints brought by members of the public; and strengthening police transparency, performance, and accountability. The main constants of this type of the system are video format, resolution, frames rate, and audio quality. This system records the video in .mp4 format with 1080p resolution and 30 frames per second. One more important aspect to while designing this system is amount of power the system requires as battery management becomes very critical. The main design challenges are Size of the Video, Audio for the video. Combining both audio and video and saving it in .mp4 format, Battery, size that is required for 8 hours of continuous recording, Security. For prototyping this system is implemented using Raspberry Pi model B.
Fitness and Independence after SCI: Defining Meaningful Change and Thresholds
2016-10-01
STATEMENT Approved for Public Release; Distribution Unlimited 13. SUPPLEMENTARY NOTES 14. ABSTRACT Quality of life after SCI/D is depends more on...determine if low fitness is limiting transfer ability. 15. SUBJECT TERMS Spinal Cord Injury, Fitness, Independence, Quality of Life 16. SECURITY... quality of life . Examples include: • data or databases; • physical collections; • audio or video products; • software; • models; • educational aids
Stewart, W R; Ramsey, M W; Jones, C J
1994-08-01
A system for the measurement of arterial pulse wave velocity is described. A personal computer (PC) plug-in transputer board is used to process the audio signals from two pocket Doppler ultrasound units. The transputer is used to provide a set of bandpass digital filters on two channels. The times of excursion of power through thresholds in each filter are recorded and used to estimate the onset of systolic flow. The system does not require an additional spectrum analyser and can work in real time. The transputer architecture provides for easy integration into any wider physiological measurement system.
Sinusoidal Analysis-Synthesis of Audio Using Perceptual Criteria
NASA Astrophysics Data System (ADS)
Painter, Ted; Spanias, Andreas
2003-12-01
This paper presents a new method for the selection of sinusoidal components for use in compact representations of narrowband audio. The method consists of ranking and selecting the most perceptually relevant sinusoids. The idea behind the method is to maximize the matching between the auditory excitation pattern associated with the original signal and the corresponding auditory excitation pattern associated with the modeled signal that is being represented by a small set of sinusoidal parameters. The proposed component-selection methodology is shown to outperform the maximum signal-to-mask ratio selection strategy in terms of subjective quality.
Expert models and modeling processes associated with a computer-modeling tool
NASA Astrophysics Data System (ADS)
Zhang, Baohui; Liu, Xiufeng; Krajcik, Joseph S.
2006-07-01
Holding the premise that the development of expertise is a continuous process, this study concerns expert models and modeling processes associated with a modeling tool called Model-It. Five advanced Ph.D. students in environmental engineering and public health used Model-It to create and test models of water quality. Using think aloud technique and video recording, we captured their computer screen modeling activities and thinking processes. We also interviewed them the day following their modeling sessions to further probe the rationale of their modeling practices. We analyzed both the audio-video transcripts and the experts' models. We found the experts' modeling processes followed the linear sequence built in the modeling program with few instances of moving back and forth. They specified their goals up front and spent a long time thinking through an entire model before acting. They specified relationships with accurate and convincing evidence. Factors (i.e., variables) in expert models were clustered, and represented by specialized technical terms. Based on the above findings, we made suggestions for improving model-based science teaching and learning using Model-It.
Kelly, Christine A.; Hewett, Paul C.; Mensch, Barbara S.; Rankin, Johanna; Nsobya, Sam; Kalibala, Sam; Kakande, Pamela
2015-01-01
Understanding the transmission dynamics of HIV and other sexually transmitted infections is critically dependent on accurate behavioral data. This paper investigates the effect of questionnaire delivery mode on the quality of sexual behavior reporting in a survey conducted in Kampala in 2010 among 18–24 year old females using the women’s instrument of the 2006 Uganda Demographic and Health Survey. We compare the reported prevalence of five sexual outcomes across three interview modes: traditional face-to-face interview (FTFI) in which question rewording was permitted, FTFI administered via computer-assisted personal interviewing (CAPI) in which questions were read as written, and audio computer-assisted self-interviewing (ACASI). We then assess the validity of the data by evaluating reporting of sexual experience against three biological markers. Results suggest that ACASI elicits higher reporting of some key indicators than face-to-face interviews, but self-reports from all interview methods were subject to validity concerns when compared with biomarker data. The paper highlights the important role biomarkers play in sexual behavior research. PMID:24615574
Lee, Jung Ae; Kim, Chul Yong; Yang, Dae Sik; Yoon, Won Sup; Park, Young Je; Lee, Suk; Kim, Young Bum
2014-01-01
To investigate the effectiveness of respiratory guidance system in 4-dimensional computed tomography (4 DCT) based respiratory-gated radiation therapy (RGRT) by comparing respiratory signals and dosimetric analysis of treatment plans. The respiratory amplitude and period of the free, the audio device-guided, and the complex system-guided breathing were evaluated in eleven patients with lung or liver cancers. The dosimetric parameters were assessed by comparing free breathing CT plan and 4 DCT-based 30-70% maximal intensity projection (MIP) plan. The use of complex system-guided breathing showed significantly less variation in respiratory amplitude and period compared to the free or audio-guided breathing regarding the root mean square errors (RMSE) of full inspiration (P = 0.031), full expiration (P = 0.007), and period (P = 0.007). The dosimetric parameters including V(5 Gy), V(10 Gy), V(20 Gy), V(30 Gy), V(40 Gy), and V(50 Gy) of normal liver or lung in 4 DCT MIP plan were superior over free breathing CT plan. The reproducibility and regularity of respiratory amplitude and period were significantly improved with the complex system-guided breathing compared to the free or the audio-guided breathing. In addition, the treatment plan based on the 4D CT-based MIP images acquired with the complex system guided breathing showed better normal tissue sparing than that on the free breathing CT.
Audio-Enhanced Tablet Computers to Assess Children's Food Frequency From Migrant Farmworker Mothers.
Kilanowski, Jill F; Trapl, Erika S; Kofron, Ryan M
2013-06-01
This study sought to improve data collection in children's food frequency surveys for non-English speaking immigrant/migrant farmworker mothers using audio-enhanced tablet computers (ATCs). We hypothesized that by using technological adaptations, we would be able to improve data capture and therefore reduce lost surveys. This Food Frequency Questionnaire (FFQ), a paper-based dietary assessment tool, was adapted for ATCs and assessed consumption of 66 food items asking 3 questions for each food item: frequency, quantity of consumption, and serving size. The tablet-based survey was audio enhanced with each question "read" to participants, accompanied by food item images, together with an embedded short instructional video. Results indicated that respondents were able to complete the 198 questions from the 66 food item FFQ on ATCs in approximately 23 minutes. Compared with paper-based FFQs, ATC-based FFQs had less missing data. Despite overall reductions in missing data by use of ATCs, respondents still appeared to have difficulty with question 2 of the FFQ. Ability to score the FFQ was dependent on what sections missing data were located. Unlike the paper-based FFQs, no ATC-based FFQs were unscored due to amount or location of missing data. An ATC-based FFQ was feasible and increased ability to score this survey on children's food patterns from migrant farmworker mothers. This adapted technology may serve as an exemplar for other non-English speaking immigrant populations.
Utterance independent bimodal emotion recognition in spontaneous communication
NASA Astrophysics Data System (ADS)
Tao, Jianhua; Pan, Shifeng; Yang, Minghao; Li, Ya; Mu, Kaihui; Che, Jianfeng
2011-12-01
Emotion expressions sometimes are mixed with the utterance expression in spontaneous face-to-face communication, which makes difficulties for emotion recognition. This article introduces the methods of reducing the utterance influences in visual parameters for the audio-visual-based emotion recognition. The audio and visual channels are first combined under a Multistream Hidden Markov Model (MHMM). Then, the utterance reduction is finished by finding the residual between the real visual parameters and the outputs of the utterance related visual parameters. This article introduces the Fused Hidden Markov Model Inversion method which is trained in the neutral expressed audio-visual corpus to solve the problem. To reduce the computing complexity the inversion model is further simplified to a Gaussian Mixture Model (GMM) mapping. Compared with traditional bimodal emotion recognition methods (e.g., SVM, CART, Boosting), the utterance reduction method can give better results of emotion recognition. The experiments also show the effectiveness of our emotion recognition system when it was used in a live environment.
Development of SIR-C Ground Calibration Equipment
NASA Technical Reports Server (NTRS)
Freeman, A.; Azeem, M.; Haub, D.; Sarabandi, K.
1993-01-01
SIR-C/X-SAR is currently scheduled for launch in April 1994. SIR-C is an L-Band and C-Band, multi-polarization spaceborne SAR system developed by NASA/JPL. X- SAR is an X-Band SAR system developed by DARA/ASI. One of the problems involved in calibrating the SIR-C instrument is to make sure that the horizontal (H) and vertical (V) polarized beams are aligned in the azimuth direction, i.e.. that they are pointing in the same direction. This is important if the polarimetric performance specifications for the system are to be met. To solve this problem, we have designed and built a prototype of a low-cost ground receiver capable of recording received power from two antennas, one H-polarized, the other V-polarized. The two signals are mixed to audio then recorded on the left and right stereo channels of a standard audio cassette player. The audio cassette recording can then be played back directly into a Macintosh computer, where it is digitized. Analysis of.
The role of laryngoscopy in the diagnosis of spasmodic dysphonia.
Daraei, Pedram; Villari, Craig R; Rubin, Adam D; Hillel, Alexander T; Hapner, Edie R; Klein, Adam M; Johns, Michael M
2014-03-01
Spasmodic dysphonia (SD) can be difficult to diagnose, and patients often see multiple physicians for many years before diagnosis. Improving the speed of diagnosis for individuals with SD may decrease the time to treatment and improve patient quality of life more quickly. To assess whether the diagnosis of SD can be accurately predicted through auditory cues alone without the assistance of visual cues offered by laryngoscopic examination. Single-masked, case-control study at a specialized referral center that included patients who underwent laryngoscopic examination as part of a multidisciplinary workup for dysphonia. Twenty-two patients were selected in total: 10 with SD, 5 with vocal tremor, and 7 controls without SD or vocal tremor. The laryngoscopic examination was recorded, deidentified, and edited to make 3 media clips for each patient: video alone, audio alone, and combined video and audio. These clips were randomized and presented to 3 fellowship-trained laryngologist raters (A.D.R., A.T.H., and A.M.K.), who established the most probable diagnosis for each clip. Intrarater and interrater reliability were evaluated using repeat clips incorporated in the presentations. We measured diagnostic accuracy for video-only, audio-only, and combined multimedia clips. These measures were established before data collection. Data analysis was accomplished with analysis of variance and Tukey honestly significant differences. Of patients with SD, diagnostic accuracy was 10%, 73%, and 73% for video-only, audio-only, and combined, respectively (P < .001, df = 2). Of patients with vocal tremor, diagnostic accuracy was 93%, 73%, and 100% for video-only, audio-only, and combined, respectively (P = .05, df = 2). Of the controls, diagnostic accuracy was 81%, 19%, and 62% for video-only, audio-only, and combined, respectively (P < .001, df = 2). The diagnosis of SD during examination is based primarily on auditory cues. Viewing combined audio and video clips afforded no change in diagnostic accuracy compared with audio alone. Laryngoscopy serves an important role in the diagnosis of SD by excluding other pathologic causes and identifying vocal tremor.
ERIC Educational Resources Information Center
Lowman, Charles E.
A guide to the technology of magnetic recorders used in such fields as audio recording, broadcast and closed-circuit television, instrumentation recording, and computer data systems is presented. Included are discussions of applications, advantages, and limitations of magnetic recording, its basic principles and theory of operation, and its…
Emerging Organizational Electronic Communication Technologies: A Selected Review of the Literature.
ERIC Educational Resources Information Center
Hellweg, Susan A.; And Others
A selective review of research dealing with emerging organizational electronic communication technologies from the communication, management, and organizational psychology literature was divided into four categories: word processing, electronic mail, computer conferencing, and teleconferencing (audio/video). The analysis was directed specifically…
ERIC Educational Resources Information Center
Abbott, George L.; And Others
1987-01-01
This special feature focuses on recent developments in optical disk technology. Nine articles discuss current trends, large scale image processing, data structures for optical disks, the use of computer simulators to create optical disks, videodisk use in training, interactive audio video systems, impacts on federal information policy, and…
76 FR 75522 - Submission for OMB Review; Comment Request
Federal Register 2010, 2011, 2012, 2013, 2014
2011-12-02
... Recorded Interviewing (CARI) technology field test using the 2012 Survey of Income and Program Participation Event History Calendar (SIPP-EHC) Field Test questionnaire. Computer Audio Recorded Interviewing... the technology. Other tests have also been conducted on non-voluntary surveys and proved promising...
ERIC Educational Resources Information Center
Hanemann, Ulrike, Ed.
2014-01-01
Different technologies have been used for decades to support adult education and learning. These include radio, television and audio and video cassettes. More recently digital ICTs such as computers, tablets, e-books, and mobile technology have spread at great speed and also found their way into the teaching and learning of literacy and numeracy…
NASA Astrophysics Data System (ADS)
Riera-Palou, Felip; den Brinker, Albertus C.
2007-12-01
This paper introduces a new audio and speech broadband coding technique based on the combination of a pulse excitation coder and a standardized parametric coder, namely, MPEG-4 high-quality parametric coder. After presenting a series of enhancements to regular pulse excitation (RPE) to make it suitable for the modeling of broadband signals, it is shown how pulse and parametric codings complement each other and how they can be merged to yield a layered bit stream scalable coder able to operate at different points in the quality bit rate plane. The performance of the proposed coder is evaluated in a listening test. The major result is that the extra functionality of the bit stream scalability does not come at the price of a reduced performance since the coder is competitive with standardized coders (MP3, AAC, SSC).
Establishing a gold standard for manual cough counting: video versus digital audio recordings
Smith, Jaclyn A; Earis, John E; Woodcock, Ashley A
2006-01-01
Background Manual cough counting is time-consuming and laborious; however it is the standard to which automated cough monitoring devices must be compared. We have compared manual cough counting from video recordings with manual cough counting from digital audio recordings. Methods We studied 8 patients with chronic cough, overnight in laboratory conditions (diagnoses were 5 asthma, 1 rhinitis, 1 gastro-oesophageal reflux disease and 1 idiopathic cough). Coughs were recorded simultaneously using a video camera with infrared lighting and digital sound recording. The numbers of coughs in each 8 hour recording were counted manually, by a trained observer, in real time from the video recordings and using audio-editing software from the digital sound recordings. Results The median cough frequency was 17.8 (IQR 5.9–28.7) cough sounds per hour in the video recordings and 17.7 (6.0–29.4) coughs per hour in the digital sound recordings. There was excellent agreement between the video and digital audio cough rates; mean difference of -0.3 coughs per hour (SD ± 0.6), 95% limits of agreement -1.5 to +0.9 coughs per hour. Video recordings had poorer sound quality even in controlled conditions and can only be analysed in real time (8 hours per recording). Digital sound recordings required 2–4 hours of analysis per recording. Conclusion Manual counting of cough sounds from digital audio recordings has excellent agreement with simultaneous video recordings in laboratory conditions. We suggest that ambulatory digital audio recording is therefore ideal for validating future cough monitoring devices, as this as this can be performed in the patients own environment. PMID:16887019
Stochastic Packet Loss Model to Evaluate QoE Impairments
NASA Astrophysics Data System (ADS)
Hohlfeld, Oliver
With provisioning of broadband access for mass market—even in wireless and mobile networks—multimedia content, especially real-time streaming of high-quality audio and video, is extensively viewed and exchanged over the Internet. Quality of Experience (QoE) aspects, describing the service quality perceived by the user, is a vital factor in ensuring customer satisfaction in today's communication networks. Frameworks for accessing quality degradations in streamed video currently are investigated as a complex multi-layered research topic, involving network traffic load, codec functions and measures of user perception of video quality.
Deamant, C.; Smith, J.; Garcia, D.; Angulo, F.
2015-01-01
Summary Background Routine implementation of instruments to capture patient-reported outcomes could guide clinical practice and facilitate health services research. Audio interviews facilitate self-interviews across literacy levels. Objectives To evaluate time burden for patients, and factors associated with response times for an audio computer-assisted self interview (ACASI) system integrated into the clinical workflow. Methods We developed an ACASI system, integrated with a research data warehouse. Instruments for symptom burden, self-reported health, depression screening, tobacco use, and patient satisfaction were administered through touch-screen monitors in the general medicine clinic at the Cook County Health & Hospitals System during April 8, 2011-July 27, 2012. We performed a cross-sectional study to evaluate the mean time burden per item and for each module of instruments; we evaluated factors associated with longer response latency. Results Among 1,670 interviews, the mean per-question response time was 18.4 [SD, 6.1] seconds. By multivariable analysis, age was most strongly associated with prolonged response time and increased per decade compared to < 50 years as follows (additional seconds per question; 95% CI): 50–59 years (1.4; 0.7 to 2.1 seconds); 60–69 (3.4; 2.6 to 4.1); 70–79 (5.1; 4.0 to 6.1); and 80–89 (5.5; 4.1 to 7.0). Response times also were longer for Spanish language (3.9; 2.9 to 4.9); no home computer use (3.3; 2.8 to 3.9); and, low mental self-reported health (0.6; 0.0 to 1.1). However, most interviews were completed within 10 minutes. Conclusions An ACASI software system can be included in a patient visit and adds minimal time burden. The burden was greatest for older patients, interviews in Spanish, and for those with less computer exposure. A patient’s self-reported health had minimal impact on response times. PMID:25848420
Trick, W E; Deamant, C; Smith, J; Garcia, D; Angulo, F
2015-01-01
Routine implementation of instruments to capture patient-reported outcomes could guide clinical practice and facilitate health services research. Audio interviews facilitate self-interviews across literacy levels. To evaluate time burden for patients, and factors associated with response times for an audio computer-assisted self interview (ACASI) system integrated into the clinical workflow. We developed an ACASI system, integrated with a research data warehouse. Instruments for symptom burden, self-reported health, depression screening, tobacco use, and patient satisfaction were administered through touch-screen monitors in the general medicine clinic at the Cook County Health & Hospitals System during April 8, 2011-July 27, 2012. We performed a cross-sectional study to evaluate the mean time burden per item and for each module of instruments; we evaluated factors associated with longer response latency. Among 1,670 interviews, the mean per-question response time was 18.4 [SD, 6.1] seconds. By multivariable analysis, age was most strongly associated with prolonged response time and increased per decade compared to < 50 years as follows (additional seconds per question; 95% CI): 50-59 years (1.4; 0.7 to 2.1 seconds); 60-69 (3.4; 2.6 to 4.1); 70-79 (5.1; 4.0 to 6.1); and 80-89 (5.5; 4.1 to 7.0). Response times also were longer for Spanish language (3.9; 2.9 to 4.9); no home computer use (3.3; 2.8 to 3.9); and, low mental self-reported health (0.6; 0.0 to 1.1). However, most interviews were completed within 10 minutes. An ACASI software system can be included in a patient visit and adds minimal time burden. The burden was greatest for older patients, interviews in Spanish, and for those with less computer exposure. A patient's self-reported health had minimal impact on response times.
Lor, Maichou; Bowers, Barbara J
2017-08-01
Many older adult immigrants in the US, including Hmong older adults, have limited English proficiency (LEP), and cannot read or have difficulty reading even in their first language (non-literate [NL]). Little has been done to identify feasible data collection approaches to enable inclusion of LEP or NL populations in research, limiting knowledge about their health. This study's purpose was to test the feasibility of culturally and linguistically adapted audio computer-assisted self-interviewing (ACASI) with color-labeled response categories and helper assistance (ACASI-H) for collection of health data with Hmong older adults. Thirty dyads (older adult and a helper) completed an ACASI-H survey with 13 health questions and a face-to-face debriefing interview. ACASI-H survey completion was video-recorded and reviewed with participants. Video review and debriefing interviews were audio-recorded and transcribed. Directed and conventional content analyses were used to analyze the interviews. All respondents reported that ACASI-H survey questions were consistent with their health experience. They lacked computer experience and found ACASI-H's interface user-friendly. All used the pre-recorded Hmong oral translation except for one, whose helper provided translation. Some Hmong older adults struggled with the color labeling at first, but helpers guided them to use the colors correctly. All dyads liked the color-labeled response categories and confirmed that a helper was necessary during the survey process. Findings support use of oral survey question administration with a technologically competent helper and color-labeled response categories when engaging LEP older adults in health-related data collection. © 2017 Wiley Periodicals, Inc. © 2017 Wiley Periodicals, Inc.
On the Acoustics of Emotion in Audio: What Speech, Music, and Sound have in Common.
Weninger, Felix; Eyben, Florian; Schuller, Björn W; Mortillaro, Marcello; Scherer, Klaus R
2013-01-01
WITHOUT DOUBT, THERE IS EMOTIONAL INFORMATION IN ALMOST ANY KIND OF SOUND RECEIVED BY HUMANS EVERY DAY: be it the affective state of a person transmitted by means of speech; the emotion intended by a composer while writing a musical piece, or conveyed by a musician while performing it; or the affective state connected to an acoustic event occurring in the environment, in the soundtrack of a movie, or in a radio play. In the field of affective computing, there is currently some loosely connected research concerning either of these phenomena, but a holistic computational model of affect in sound is still lacking. In turn, for tomorrow's pervasive technical systems, including affective companions and robots, it is expected to be highly beneficial to understand the affective dimensions of "the sound that something makes," in order to evaluate the system's auditory environment and its own audio output. This article aims at a first step toward a holistic computational model: starting from standard acoustic feature extraction schemes in the domains of speech, music, and sound analysis, we interpret the worth of individual features across these three domains, considering four audio databases with observer annotations in the arousal and valence dimensions. In the results, we find that by selection of appropriate descriptors, cross-domain arousal, and valence regression is feasible achieving significant correlations with the observer annotations of up to 0.78 for arousal (training on sound and testing on enacted speech) and 0.60 for valence (training on enacted speech and testing on music). The high degree of cross-domain consistency in encoding the two main dimensions of affect may be attributable to the co-evolution of speech and music from multimodal affect bursts, including the integration of nature sounds for expressive effects.
Detection of goal events in soccer videos
NASA Astrophysics Data System (ADS)
Kim, Hyoung-Gook; Roeber, Steffen; Samour, Amjad; Sikora, Thomas
2005-01-01
In this paper, we present an automatic extraction of goal events in soccer videos by using audio track features alone without relying on expensive-to-compute video track features. The extracted goal events can be used for high-level indexing and selective browsing of soccer videos. The detection of soccer video highlights using audio contents comprises three steps: 1) extraction of audio features from a video sequence, 2) event candidate detection of highlight events based on the information provided by the feature extraction Methods and the Hidden Markov Model (HMM), 3) goal event selection to finally determine the video intervals to be included in the summary. For this purpose we compared the performance of the well known Mel-scale Frequency Cepstral Coefficients (MFCC) feature extraction method vs. MPEG-7 Audio Spectrum Projection feature (ASP) extraction method based on three different decomposition methods namely Principal Component Analysis( PCA), Independent Component Analysis (ICA) and Non-Negative Matrix Factorization (NMF). To evaluate our system we collected five soccer game videos from various sources. In total we have seven hours of soccer games consisting of eight gigabytes of data. One of five soccer games is used as the training data (e.g., announcers' excited speech, audience ambient speech noise, audience clapping, environmental sounds). Our goal event detection results are encouraging.
The Personal Hearing System—A Software Hearing Aid for a Personal Communication System
NASA Astrophysics Data System (ADS)
Grimm, Giso; Guilmin, Gwénaël; Poppen, Frank; Vlaming, Marcel S. M. G.; Hohmann, Volker
2009-12-01
A concept and architecture of a personal communication system (PCS) is introduced that integrates audio communication and hearing support for the elderly and hearing-impaired through a personal hearing system (PHS). The concept envisions a central processor connected to audio headsets via a wireless body area network (WBAN). To demonstrate the concept, a prototype PCS is presented that is implemented on a netbook computer with a dedicated audio interface in combination with a mobile phone. The prototype can be used for field-testing possible applications and to reveal possibilities and limitations of the concept of integrating hearing support in consumer audio communication devices. It is shown that the prototype PCS can integrate hearing aid functionality, telephony, public announcement systems, and home entertainment. An exemplary binaural speech enhancement scheme that represents a large class of possible PHS processing schemes is shown to be compatible with the general concept. However, an analysis of hardware and software architectures shows that the implementation of a PCS on future advanced cell phone-like devices is challenging. Because of limitations in processing power, recoding of prototype implementations into fixed point arithmetic will be required and WBAN performance is still a limiting factor in terms of data rate and delay.
ERIC Educational Resources Information Center
Davies, Maree; Kiemer, Katharina; Meissel, Kane
2017-01-01
This study used the Quality Talk and dialogic teaching approach with a group of secondary school teachers (N = 7) to train their facilitation of dialogical discussions by small groups of students. The study used video and audio analysis to assess the teachers' observable behaviours during these discussions, before and after professional…
Transmission of live laparoscopic surgery over the Internet2.
Damore, L J; Johnson, J A; Dixon, R S; Iverson, M A; Ellison, E C; Melvin, W S
1999-11-01
Video broadcasting of surgical procedures is an important tool for education, training, and consultation. Current video conferencing systems are expensive and time-consuming and require preplanning. Real-time Internet video is known for its poor quality and relies on the equipment and the speed of the connection. The Internet2, a new high-speed (up to 2,048 Mbps), large bandwidth data network presently connects more than 100 universities and corporations. We have successfully used the Internet2 to broadcast the first real-time, high-quality audio/video program from a live laparoscopic operation to distant points. Video output of the laparoscopic camera and audio from a wireless microphone were broadcast to distant sites using a proprietary, PC-based implementation of H.320 video conferencing over a TCP/IP network connected to the Internet2. The receiving sites participated in two-way, real-time video and audio communications and graded the quality of the signal they received. On August 25, 1998, a laparoscopic Nissen fundoplication was transmitted to Internet2 stations in Colorado, Pennsylvania, and to an Internet station in New York. On September 28 and 29, 1998, we broadcast laparoscopic operations throughout both days to the Internet2 Fall Conference in San Francisco, California. Most recently, on February 24, 1999, we transmitted a laparoscopic Heller myotomy to the Abilene Network Launch Event in Washington, DC. The Internet2 is currently able to provide the bandwidth needed for a turn-key video conferencing system with high-resolution, real-time transmission. The system could be used for a variety of teaching and educational programs for experienced surgeons, residents, and medical students.
Open University Environmental Education and Training.
ERIC Educational Resources Information Center
Blackmore, Christine
1996-01-01
Describes the approach to environmental education courses at Open University. Includes broad course content, team teaching approach, and philosophy of reorienting education towards sustainable development. Course material for open learning includes self-contained study packs as well as course texts, video, audio, and computer software. Nonformal…
Code of Federal Regulations, 2014 CFR
2014-01-01
... Photographic and photocopying equipment manufacturing. 3341 Computer and peripheral equipment manufacturing. 33422 Radio and television broadcasting and wireless communications equipment manufacturing. 33429 Other communications equipment manufacturing. 3343 Audio and video equipment manufacturing. 334412 Bare printed circuit...
ERIC Educational Resources Information Center
Fernandez, Kim
2010-01-01
With more and more people attached to their computers, it's no wonder that publications are increasingly going online. Magazines are either supplementing their print content with online bonus information, such as extended features, photos, audio files, or videos, or looking to ditch the printing presses entirely to focus on all-electronic…
Delivery Systems for Distance Education. ERIC Digest.
ERIC Educational Resources Information Center
Schamber, Linda
This ERIC digest provides a brief overview of the video, audio, and computer technologies that are currently used to deliver instruction for distance education programs. The video systems described include videoconferencing, low-power television (LPTV), closed-circuit television (CCTV), instructional fixed television service (ITFS), and cable…
New Horizons in Adult Education. Volumes 3-7. 1989-1993.
ERIC Educational Resources Information Center
New Horizons in Adult Education, 1993
1993-01-01
Volume 3 includes the following: "Comparison of Computer and Audio Teleconferencing" (Norman Coombs); "Intellectual Suppression" [book review] (Roger Boshier). Contents of volume 4 are as follows: "Straight Time and Standard Brand Adult Education" (John Ohliger); "Comparison of Folk High Schools in Denmark, and…
34 CFR 388.22 - What priorities does the Secretary consider in making an award?
Code of Federal Regulations, 2011 CFR
2011-07-01
... education methods, such as interactive audio, video, computer technologies, or existing telecommunications... training materials and practices. The proposed project demonstrates an effective plan to develop and... programs by other State vocational rehabilitation units. (2) Distance education. The proposed project...
34 CFR 388.22 - What priorities does the Secretary consider in making an award?
Code of Federal Regulations, 2010 CFR
2010-07-01
... education methods, such as interactive audio, video, computer technologies, or existing telecommunications... training materials and practices. The proposed project demonstrates an effective plan to develop and... programs by other State vocational rehabilitation units. (2) Distance education. The proposed project...
CWRUnet--Case History of a Campus-Wide Fiber-to-the-Desktop Network.
ERIC Educational Resources Information Center
Neff, Raymond K.; Haigh, Peter J.
1992-01-01
This article describes the development at Case Western Reserve University of an all-fiber optic communications network linking 7,300 outlets (faculty offices, student residences, classrooms, libraries, and laboratories) with computer data, television, audio, facsimile, and image information services. (Author/DB)
34 CFR 388.22 - What priorities does the Secretary consider in making an award?
Code of Federal Regulations, 2014 CFR
2014-07-01
... education methods, such as interactive audio, video, computer technologies, or existing telecommunications... training materials and practices. The proposed project demonstrates an effective plan to develop and... programs by other State vocational rehabilitation units. (2) Distance education. The proposed project...
34 CFR 388.22 - What priorities does the Secretary consider in making an award?
Code of Federal Regulations, 2013 CFR
2013-07-01
... education methods, such as interactive audio, video, computer technologies, or existing telecommunications... training materials and practices. The proposed project demonstrates an effective plan to develop and... programs by other State vocational rehabilitation units. (2) Distance education. The proposed project...
34 CFR 388.22 - What priorities does the Secretary consider in making an award?
Code of Federal Regulations, 2012 CFR
2012-07-01
... education methods, such as interactive audio, video, computer technologies, or existing telecommunications... training materials and practices. The proposed project demonstrates an effective plan to develop and... programs by other State vocational rehabilitation units. (2) Distance education. The proposed project...
Audio Classification in Speech and Music: A Comparison between a Statistical and a Neural Approach
NASA Astrophysics Data System (ADS)
Bugatti, Alessandro; Flammini, Alessandra; Migliorati, Pierangelo
2002-12-01
We focus the attention on the problem of audio classification in speech and music for multimedia applications. In particular, we present a comparison between two different techniques for speech/music discrimination. The first method is based on Zero crossing rate and Bayesian classification. It is very simple from a computational point of view, and gives good results in case of pure music or speech. The simulation results show that some performance degradation arises when the music segment contains also some speech superimposed on music, or strong rhythmic components. To overcome these problems, we propose a second method, that uses more features, and is based on neural networks (specifically a multi-layer Perceptron). In this case we obtain better performance, at the expense of a limited growth in the computational complexity. In practice, the proposed neural network is simple to be implemented if a suitable polynomial is used as the activation function, and a real-time implementation is possible even if low-cost embedded systems are used.
Folan, Alyce; Barclay, Linda; Cooper, Cathy; Robinson, Merren
2015-01-01
Assistive technology for computer access can be used to facilitate people with a spinal cord injury to utilize mainstream computer applications, thereby enabling participation in a variety of meaningful occupations. The aim of this study was to gain an understanding of the experiences of clients with tetraplegia trialing assistive technologies for computer access during different stages in a public rehabilitation service. In order to explore the experiences of clients with tetraplegia trialing assistive technologies for computer use, qualitative methodology was selected. Data were collected from seven participants using semi-structured interviews, which were audio-taped, transcribed and analyzed thematically. Three main themes were identified. These were: getting back into life, assisting in adjusting to injury and learning new skills. The findings from this study demonstrated that people with tetraplegia can be assisted to return to previous life roles or engage in new roles, through developing skills in the use of assistive technology for computer access. Being able to use computers for meaningful activities contributed to the participants gaining an enhanced sense of self-efficacy, and thereby quality of life. Implications for Rehabilitation Findings from this pilot study indicate that people with tetraplegia can be assisted to return to previous life roles, and develop new roles that have meaning to them through the use of assistive technologies for computer use. Being able to use the internet to socialize, and complete daily tasks, contributed to the participants gaining a sense of control over their lives. Early introduction to assistive technology is important to ensure sufficient time for newly injured people to feel comfortable enough with the assistive technology to use the computers productively by the time of discharge. Further research into this important and expanding area is indicated.
Real-time implementation of an interactive jazz accompaniment system
NASA Astrophysics Data System (ADS)
Deshpande, Nikhil
Modern computational algorithms and digital signal processing (DSP) are able to combine with human performers without forced or predetermined structure in order to create dynamic and real-time accompaniment systems. With modern computing power and intelligent algorithm layout and design, it is possible to achieve more detailed auditory analysis of live music. Using this information, computer code can follow and predict how a human's musical performance evolves, and use this to react in a musical manner. This project builds a real-time accompaniment system to perform together with live musicians, with a focus on live jazz performance and improvisation. The system utilizes a new polyphonic pitch detector and embeds it in an Ableton Live system - combined with Max for Live - to perform elements of audio analysis, generation, and triggering. The system also relies on tension curves and information rate calculations from the Creative Artificially Intuitive and Reasoning Agent (CAIRA) system to help understand and predict human improvisation. These metrics are vital to the core system and allow for extrapolated audio analysis. The system is able to react dynamically to a human performer, and can successfully accompany the human as an entire rhythm section.
Publication of science data on CD-ROM: A guide and example
NASA Technical Reports Server (NTRS)
Angelici, Gary; Skiles, J. W.
1993-01-01
CD-ROM (Compact Disk-Read Only Memory) is becoming the standard media not only in audio recording, but also in the publication of data and information accessible on many computer platforms. Little has been written about the complicated process involved in creating easy-to-use, high quality, and useful CD-ROM's containing scientific data. This document is a manual designed to aid those who are responsible for the publication of scientific data on CD-ROM. All aspects and steps of the procedure are covered, from feasibility assessment through disk design, data preparation, disc mastering, and CD-ROM distribution. General advice and actual examples are based on lessons learned from the publication of scientific data for an interdisciplinary field experiment. Appendices include actual files from a CD-ROM, a purchase request for CD-ROM mastering services, and the disk art for the first disk published for the project.
Virtual environment display for a 3D audio room simulation
NASA Astrophysics Data System (ADS)
Chapin, William L.; Foster, Scott
1992-06-01
Recent developments in virtual 3D audio and synthetic aural environments have produced a complex acoustical room simulation. The acoustical simulation models a room with walls, ceiling, and floor of selected sound reflecting/absorbing characteristics and unlimited independent localizable sound sources. This non-visual acoustic simulation, implemented with 4 audio ConvolvotronsTM by Crystal River Engineering and coupled to the listener with a Poihemus IsotrakTM, tracking the listener's head position and orientation, and stereo headphones returning binaural sound, is quite compelling to most listeners with eyes closed. This immersive effect should be reinforced when properly integrated into a full, multi-sensory virtual environment presentation. This paper discusses the design of an interactive, visual virtual environment, complementing the acoustic model and specified to: 1) allow the listener to freely move about the space, a room of manipulable size, shape, and audio character, while interactively relocating the sound sources; 2) reinforce the listener's feeling of telepresence into the acoustical environment with visual and proprioceptive sensations; 3) enhance the audio with the graphic and interactive components, rather than overwhelm or reduce it; and 4) serve as a research testbed and technology transfer demonstration. The hardware/software design of two demonstration systems, one installed and one portable, are discussed through the development of four iterative configurations. The installed system implements a head-coupled, wide-angle, stereo-optic tracker/viewer and multi-computer simulation control. The portable demonstration system implements a head-mounted wide-angle, stereo-optic display, separate head and pointer electro-magnetic position trackers, a heterogeneous parallel graphics processing system, and object oriented C++ program code.
ERIC Educational Resources Information Center
Sayre, Scott Alan
The purpose of this study was to develop and validate a computer-based system that would allow interactive video developers to integrate and manage the design components prior to production. These components of an interactive video (IVD) program include visual information in a variety of formats, audio information, and instructional techniques,…
NASA Astrophysics Data System (ADS)
Varner, R. K.; Palace, M. W.; Lennartz, J. M.; Crill, P. M.; Wik, M.; Amante, J.; Dorich, C.; Harden, J. W.; Ewing, S. A.; Turetsky, M. R.
2011-12-01
Knowledge of the magnitude and frequency of methane release through ebullition (bubbling) in water saturated ecosystems such as bogs, fens and lakes is important to both the atmospheric and ecosystems science community. The controls on episodic bubble releases must be identified in order to understand the response of these ecosystems to future climate forcing. We have developed and field tested an inexpensive array of sampling/monitoring instruments to identify the frequency and magnitude of bubbling events which allows us to correlate bubble data with potential drivers such as changes in hydrostatic pressure, wind and temperature. A prototype ebullition sensor has been developed and field tested at Sallie's Fen in New Hampshire, USA. The instrument consists of a nested, inverted funnel design with a hydrophone for detecting bubbles rising through the peat, that hit the microphone. The design also offers a way to sample the gases collected from the funnels to determine the concentration of CH4. Laboratory calibration of the instrument resulted in an equation that relates frequency of bubbles hitting the microphone with bubble volume. After calibration in the laboratory, the prototype was deployed in Sallie's Fen in late August 2010. An additional four instruments were deployed the following month. Audio data was recorded continuously using a digital audio recorder attached to two ebullition sensors. Audio was recorded as an mp3 compressed audio file at a sample rate of 160 kbits/sec. Using this format and stereo input, allowing for two sensors to be recorded with each device, we were able to record continuously for 20 days. Audio was converted to uncompressed audio files for speed in computation. Audio data was processed using MATLAB, searching in 0.5 second incremental sections for specific fundamental frequencies that are related to our calibrated audio events. Time, fundamental frequency, and estimated bubble size were output to a text file for analysis in statistical software. In addition, each event was cut out of the longer audio file and placed in a directory with number of ebullition event, sensor number, and time, allowing for manual interpretation of the ebullition event. After successful laboratory and local field testing, our instruments were deployed in summer 2011 at a temperate fen (Sallie's Fen, NH, USA), a subarctic mire and lake (Stordalen, Abisko, Sweden) and two locations in subarctic Alaska (APEX Research Site, Fairbanks, AK and Innoko National Wildlife Refuge). Ebullition occurred at regular intervals. Our results indicate that this is a useful method for monitoring CH4 ebullitive flux at high temporal frequencies.
Non-Print Social Studies Materials--Elementary School Level.
ERIC Educational Resources Information Center
Lynn, Karen
Types of non-print social studies materials developed for presentation to, and use by, elementary school students are identified. "Non-print" materials include films, filmstrips, video cassettes, audio recordings, computer databases, telecommunications, and hypertext. An explanation of why elementary school students can benefit from the use of…
76 FR 4078 - Television Broadcasting Services; North Pole and Plattsburgh, NY
Federal Register 2010, 2011, 2012, 2013, 2014
2011-01-24
... FEDERAL COMMUNICATIONS COMMISSION 47 CFR Part 73 [DA 10-2443; MM Docket No. 99-238; RM-9669] Television Broadcasting Services; North Pole and Plattsburgh, NY AGENCY: Federal Communications Commission... document in accessible formats (computer diskettes, large print, audio recording, and Braille), send an e...
Code of Federal Regulations, 2011 CFR
2011-01-01
... Industry Classification System (NAICS) codes: 2007 NAICS codes 2007 NAICS industry titles 3341 Computer and peripheral equipment manufacturing. 33422 Radio and television broadcasting and wireless communications equipment manufacturing. 33429 Other communications equipment manufacturing. 3343 Audio and video equipment...
Code of Federal Regulations, 2012 CFR
2012-01-01
... Industry Classification System (NAICS) codes: 2007 NAICS codes 2007 NAICS industry titles 3341 Computer and peripheral equipment manufacturing. 33422 Radio and television broadcasting and wireless communications equipment manufacturing. 33429 Other communications equipment manufacturing. 3343 Audio and video equipment...
Code of Federal Regulations, 2013 CFR
2013-01-01
... Industry Classification System (NAICS) codes: 2007 NAICS codes 2007 NAICS industry titles 3341 Computer and peripheral equipment manufacturing. 33422 Radio and television broadcasting and wireless communications equipment manufacturing. 33429 Other communications equipment manufacturing. 3343 Audio and video equipment...
Australian DefenceScience. Volume 16, Number 1, Autumn
2008-01-01
are carried via VOIP technology, and multicast IP traffic for audio -visual communications is also supported. The SSATIN system overall is seen to...Artificial Intelligence and Soft Computing Palma de Mallorca, Spain http://iasted.com/conferences/home-628.html 1 - 3 Sep 2008 Visualisation , Imaging and
The use of multimedia and programmed teaching machines for remote sensing education
NASA Technical Reports Server (NTRS)
Ulliman, J. J.
1980-01-01
The advantages, limitations, and uses of various audio visual equipments and techniques used in various universities for individualized and group instruction in the interpretation and classification of remotely sensed data are considered as well as systems for programmed and computer-assisted instruction.
An Algorithm for Controlled Integration of Sound and Text.
ERIC Educational Resources Information Center
Wohlert, Harry S.; McCormick, Martin
1985-01-01
A serious drawback in introducing sound into computer programs for teaching foreign language speech has been the lack of an algorithm to turn off the cassette recorder immediately to keep screen text and audio in synchronization. This article describes a program which solves that problem. (SED)
Language Labs for 1990: User-Friendly, Expandable and Affordable.
ERIC Educational Resources Information Center
Wiley, Patricia Davis
1990-01-01
Describes hardware available for school laboratories used for second-language learning. Vendors and prices for equipment ranging from simple audio to computer interactive capabilities are included, portable and fixed installations are reviewed, specifications for instructor consoles and student stations are suggested, and maintenance and repair…
Digital Audio Tape: Yet Another Archival Media?
ERIC Educational Resources Information Center
Vanker, Anthony D.
1989-01-01
Provides an introduction to the technical aspects of digital audiotape and compares it to other computer storage devices such as optical data disks and magnetic tape cartridges in terms of capacity, transfer rate, and cost. The current development of digital audiotape standards is also discussed. (five references) (CLB)
Doctoral Research in Library Media; Completed and Underway 1970.
ERIC Educational Resources Information Center
Anderton, Ray L., Ed.; Mapes, Joseph L., Ed.
Doctoral theses completed and doctoral theses underway in the subject area of instructional technology are listed in this bibliography under the subtitles of audio literacy, audiovisual techniques, computers in education, library media, media training, programed instruction, projected materials, simulation and games, systems approach, television,…
ERIC Educational Resources Information Center
Calandra, Brendan; Brantley-Dias, Laurie; Yerby, Johnathan; Demir, Kadir
2018-01-01
A group of preservice science teachers edited video footage of their practice teaching to identify and isolate critical incidents. They then wrote guided reflection papers on those critical incidents using different forms of media prompts while they wrote. The authors used a counterbalanced research design to compare the quality of writing that…
The Quality and Frequency of Mother-Toddler Conflict: Links with Attachment and Temperament
ERIC Educational Resources Information Center
Laible, Deborah; Panfile, Tia; Makariev, Drika
2008-01-01
The goal of this study was to examine the links among attachment, child temperament, and the quality and frequency of mother-toddler conflict. Sixty-four mothers and children took part in a series of laboratory tasks when the child was 30 months of age and an audio-recorded home observation when the child was 36 months of age. All episodes of…
Audiovisual focus of attention and its application to Ultra High Definition video compression
NASA Astrophysics Data System (ADS)
Rerabek, Martin; Nemoto, Hiromi; Lee, Jong-Seok; Ebrahimi, Touradj
2014-02-01
Using Focus of Attention (FoA) as a perceptual process in image and video compression belongs to well-known approaches to increase coding efficiency. It has been shown that foveated coding, when compression quality varies across the image according to region of interest, is more efficient than the alternative coding, when all region are compressed in a similar way. However, widespread use of such foveated compression has been prevented due to two main conflicting causes, namely, the complexity and the efficiency of algorithms for FoA detection. One way around these is to use as much information as possible from the scene. Since most video sequences have an associated audio, and moreover, in many cases there is a correlation between the audio and the visual content, audiovisual FoA can improve efficiency of the detection algorithm while remaining of low complexity. This paper discusses a simple yet efficient audiovisual FoA algorithm based on correlation of dynamics between audio and video signal components. Results of audiovisual FoA detection algorithm are subsequently taken into account for foveated coding and compression. This approach is implemented into H.265/HEVC encoder producing a bitstream which is fully compliant to any H.265/HEVC decoder. The influence of audiovisual FoA in the perceived quality of high and ultra-high definition audiovisual sequences is explored and the amount of gain in compression efficiency is analyzed.
Analysis and enhancement of country singing
NASA Astrophysics Data System (ADS)
Lee, Matthew; Smith, Mark J. T.
2003-04-01
The study of human singing has focused extensively on the analysis of voice characteristics. At the same time, a substantial body of work has been under study aimed at modeling and synthesizing the human voice. The work on which we report brings together some key analysis and synthesis principles to create a new model for digitally improving the perceived quality of an average singing voice. The model presented employs an analysis-by-synthesis overlap-add (ABS-OLA) sinusoidal model, which in the past has been used for the analysis and synthesis of speech, in combination with a spectral model of the vocal tract. The ABS-OLA sinusoidal model for speech has been shown to be a flexible, accurate, and computationally efficient representation capable of producing a natural-sounding singing voice [E. B. George and M. J. T. Smith, Trans. Speech Audio Processing 5, 389-406 (1997)]. A spectral model infused in the ABS-OLA uses Generalized Gaussian functions to provide a simple framework which enables the precise modification of spectral characteristics while maintaining the quality and naturalness of the original voice. Furthermore, it is shown that the parameters of the new ABS-OLA can accommodate pitch corrections and vocal quality enhancements while preserving naturalness and singer identity. Examples of enhanced country singing will be presented.
Algorithms for highway-speed acoustic impact-echo evaluation of concrete bridge decks
NASA Astrophysics Data System (ADS)
Mazzeo, Brian A.; Guthrie, W. Spencer
2018-04-01
A new acoustic impact-echo testing device has been developed for detecting and mapping delaminations in concrete bridge decks at highway speeds. The apparatus produces nearly continuous acoustic excitation of concrete bridge decks through rolling mats of chains that are placed around six wheels mounted to a hinged trailer. The wheels approximately span the width of a traffic lane, and the ability to remotely lower and raise the apparatus using a winch system allows continuous data collection without stationary traffic control or exposure of personnel to traffic. Microphones near the wheels are used to record the acoustic response of the bridge deck during testing. In conjunction with the development of this new apparatus, advances in the algorithms required for data analysis were needed. This paper describes the general framework of the algorithms developed for converting differential global positioning system data and multi-channel audio data into maps that can be used in support of engineering decisions about bridge deck maintenance, rehabilitation, and replacement (MR&R). Acquisition of position and audio data is coordinated on a laptop computer through a custom graphical user interface. All of the streams of data are synchronized with the universal computer time so that audio data can be associated with interpolated position information through data post-processing. The audio segments are individually processed according to particular detection algorithms that can adapt to variations in microphone sensitivity or particular chain excitations. Features that are greater than a predetermined threshold, which is held constant throughout the analysis, are then subjected to further analysis and included in a map that shows the results of the testing. Maps of data collected on a bridge deck using the new acoustic impact-echo testing device at different speeds ranging from approximately 10 km/h to 55 km/h indicate that the collected data are reasonably repeatable. Use of the new acoustic impact-echo testing device is expected to enable more informed decisions about MR&R of concrete bridge decks.
Audio-Enhanced Tablet Computers to Assess Children’s Food Frequency From Migrant Farmworker Mothers
Kilanowski, Jill F.; Trapl, Erika S.; Kofron, Ryan M.
2014-01-01
This study sought to improve data collection in children’s food frequency surveys for non-English speaking immigrant/migrant farmworker mothers using audio-enhanced tablet computers (ATCs). We hypothesized that by using technological adaptations, we would be able to improve data capture and therefore reduce lost surveys. This Food Frequency Questionnaire (FFQ), a paper-based dietary assessment tool, was adapted for ATCs and assessed consumption of 66 food items asking 3 questions for each food item: frequency, quantity of consumption, and serving size. The tablet-based survey was audio enhanced with each question “read” to participants, accompanied by food item images, together with an embedded short instructional video. Results indicated that respondents were able to complete the 198 questions from the 66 food item FFQ on ATCs in approximately 23 minutes. Compared with paper-based FFQs, ATC-based FFQs had less missing data. Despite overall reductions in missing data by use of ATCs, respondents still appeared to have difficulty with question 2 of the FFQ. Ability to score the FFQ was dependent on what sections missing data were located. Unlike the paper-based FFQs, no ATC-based FFQs were unscored due to amount or location of missing data. An ATC-based FFQ was feasible and increased ability to score this survey on children’s food patterns from migrant farmworker mothers. This adapted technology may serve as an exemplar for other non-English speaking immigrant populations. PMID:25343004
Embedded security system for multi-modal surveillance in a railway carriage
NASA Astrophysics Data System (ADS)
Zouaoui, Rhalem; Audigier, Romaric; Ambellouis, Sébastien; Capman, François; Benhadda, Hamid; Joudrier, Stéphanie; Sodoyer, David; Lamarque, Thierry
2015-10-01
Public transport security is one of the main priorities of the public authorities when fighting against crime and terrorism. In this context, there is a great demand for autonomous systems able to detect abnormal events such as violent acts aboard passenger cars and intrusions when the train is parked at the depot. To this end, we present an innovative approach which aims at providing efficient automatic event detection by fusing video and audio analytics and reducing the false alarm rate compared to classical stand-alone video detection. The multi-modal system is composed of two microphones and one camera and integrates onboard video and audio analytics and fusion capabilities. On the one hand, for detecting intrusion, the system relies on the fusion of "unusual" audio events detection with intrusion detections from video processing. The audio analysis consists in modeling the normal ambience and detecting deviation from the trained models during testing. This unsupervised approach is based on clustering of automatically extracted segments of acoustic features and statistical Gaussian Mixture Model (GMM) modeling of each cluster. The intrusion detection is based on the three-dimensional (3D) detection and tracking of individuals in the videos. On the other hand, for violent events detection, the system fuses unsupervised and supervised audio algorithms with video event detection. The supervised audio technique detects specific events such as shouts. A GMM is used to catch the formant structure of a shout signal. Video analytics use an original approach for detecting aggressive motion by focusing on erratic motion patterns specific to violent events. As data with violent events is not easily available, a normality model with structured motions from non-violent videos is learned for one-class classification. A fusion algorithm based on Dempster-Shafer's theory analyses the asynchronous detection outputs and computes the degree of belief of each probable event.
A Prospectus for the Future Development of a Speech Lab: Hypertext Applications.
ERIC Educational Resources Information Center
Berube, David M.
This paper presents a plan for the next generation of speech laboratories which integrates technologies of modern communication in order to improve and modernize the instructional process. The paper first examines the application of intermediate technologies including audio-video recording and playback, computer assisted instruction and testing…
Read Naturally. Revised. What Works Clearinghouse Intervention Report
ERIC Educational Resources Information Center
What Works Clearinghouse, 2007
2007-01-01
"Read Naturally" is designed to improve reading fluency using a combination of books, audio-tapes, and computer software. This program includes three main strategies: repeated reading of English text for oral reading fluency development, teacher modeling of story reading, and systematic monitoring of student progress by teachers.…
Read Naturally. What Works Clearinghouse Intervention Report
ERIC Educational Resources Information Center
What Works Clearinghouse, 2006
2006-01-01
"Read Naturally" is designed to improve reading fluency using a combination of books, audio-tapes, and computer software. This program includes three main strategies: (1) repeated reading of English text for oral reading fluency development; (2) teacher modeling of story reading; and (3) systematic monitoring of student progress by…
ERIC Educational Resources Information Center
Indiana State Commission for Higher Education, Indianapolis.
A program schedule and background information for Indiana Commission for Higher Education-sponsored discussion of the use of educational technology to increase educational effeciency are presented. The four major topics of discussion to illustrate the uses and advantages/disadvantages of audio, video, and computing technologies are as follows:…
A Software Development Approach for Computer Assisted Language Learning
ERIC Educational Resources Information Center
Cushion, Steve
2005-01-01
Over the last 5 years we have developed, produced, tested, and evaluated an authoring software package to produce web-based, interactive, audio-enhanced language-learning material. That authoring package has been used to produce language-learning material in French, Spanish, German, Arabic, and Tamil. We are currently working on increasing…
A Scenario Approach to Assessment of New Communications Media.
ERIC Educational Resources Information Center
Spangler, Kathleen; And Others
In a study supported by the Charles F. Kettering Foundation, a research team developed a methodology for illustrating the effective and ineffective uses of audio, video, and computer teleconferencing by developing scenarios for eacb medium. The group first invented a general situation--a conference involving participants with global, regional, and…
Game-Based Evacuation Drill Using Augmented Reality and Head-Mounted Display
ERIC Educational Resources Information Center
Kawai, Junya; Mitsuhara, Hiroyuki; Shishibori, Masami
2016-01-01
Purpose: Evacuation drills should be more realistic and interactive. Focusing on situational and audio-visual realities and scenario-based interactivity, the authors have developed a game-based evacuation drill (GBED) system that presents augmented reality (AR) materials on tablet computers. The paper's current research purpose is to improve…
Virtual Patients in Geriatric Education
ERIC Educational Resources Information Center
Tan, Zaldy S.; Mulhausen, Paul L.; Smith, Stephen R.; Ruiz, Jorge G.
2010-01-01
The virtual patient is a case-based computer program that combines textual information with multimedia elements such as audio, graphics, and animation. It is increasingly being utilized as a teaching modality by medical educators in various fields of instruction. The inherent complexity of older patients and the shortage of geriatrics educators…
Probabilistic Graphical Models for the Analysis and Synthesis of Musical Audio
2010-11-01
Abbreviation for the names Griffiths, Engen , and McCloskey. Often used to de- note the stick-breaking distribution over infinite vectors whose elements...of state calculations by fast computing machines. Journal of Chemical Physics, 21:1087–1092, 1953. [65] R. Miotto, L. Barrington, and G. Lanckriet
Oral Conversations Online: Redefining Oral Competence in Synchronous Environments
ERIC Educational Resources Information Center
Lamy, Marie-Noelle
2004-01-01
In this article the focus is on methodology for analysing learner-learner oral conversations mediated by computers. With the increasing availability of synchronous voice-based groupware and the additional facilities offered by audio-graphic tools, language learners have opportunities for collaborating on oral tasks, supported by visual and textual…
Inferring Speaker Affect in Spoken Natural Language Communication
ERIC Educational Resources Information Center
Pon-Barry, Heather Roberta
2013-01-01
The field of spoken language processing is concerned with creating computer programs that can understand human speech and produce human-like speech. Regarding the problem of understanding human speech, there is currently growing interest in moving beyond speech recognition (the task of transcribing the words in an audio stream) and towards…
Things the Teacher of Your Media Utilization Course May Not Have Told You.
ERIC Educational Resources Information Center
Ekhaml, Leticia
1995-01-01
Discusses maintenance and safety information that may not be covered in a technology training program. Topics include computers, printers, televisions, video and audio equipment, electric roll laminators, overhead and slide projectors, equipment carts, power cords and outlets, batteries, darkrooms, barcode readers, Liquid Crystal Display units,…
77 FR 29350 - Agency Forms Undergoing Paperwork Reduction Act Review
Federal Register 2010, 2011, 2012, 2013, 2014
2012-05-17
... using audio computer assisted self-interview (ACASI). The ACASI interview includes questions about... DEPARTMENT OF HEALTH AND HUMAN SERVICES Centers for Disease Control and Prevention [30Day-12-12EL... Exit Interview 10 1 30/60 Kimberly S. Lane, Deputy Director, Office of Science Integrity, Office of the...
ERIC Educational Resources Information Center
Gooler, Dennis D., Ed.
This resource guide for community college teachers and administrators focuses on hardware and software. The following are discussed: (1) individual technologies--computer-assisted instruction, audio tape, films, filmstrips/slides, dial access, programmed instruction, learning activity packages, video cassettes, cable TV, independent learning labs,…
Home and School Technology: Wired versus Wireless.
ERIC Educational Resources Information Center
Van Horn, Royal
2001-01-01
Presents results of informal research on smart homes and appliances, structured home wiring, whole-house audio/video distribution, hybrid cable, and wireless networks. Computer network wiring is tricky to install unless all-in-one jacketed cable is used. Wireless phones help installers avoid pre-wiring problems in homes and schools. (MLH)
Visual-Auditory Integration during Speech Imitation in Autism
ERIC Educational Resources Information Center
Williams, Justin H. G.; Massaro, Dominic W.; Peel, Natalie J.; Bosseler, Alexis; Suddendorf, Thomas
2004-01-01
Children with autistic spectrum disorder (ASD) may have poor audio-visual integration, possibly reflecting dysfunctional "mirror neuron" systems which have been hypothesised to be at the core of the condition. In the present study, a computer program, utilizing speech synthesizer software and a "virtual" head (Baldi), delivered speech stimuli for…
Communication Satellites and Education in Indonesia: What Is an Appropriate Strategy?
ERIC Educational Resources Information Center
White, Peter B.; Kelabora, Lambert
1978-01-01
Advocates the use of radio and audio cassette recorders to meet the needs of the Indonesian educational system, i.e., for rural education, to widen educational opportunities, improve the quality of education, and to train people for employment. (JEG)
Direct measurement of the speed of sound using a microphone and a speaker
NASA Astrophysics Data System (ADS)
Gómez-Tejedor, José A.; Castro-Palacio, Juan C.; Monsoriu, Juan A.
2014-05-01
We present a simple and accurate experiment to obtain the speed of sound in air using a conventional speaker and a microphone connected to a computer. A free open source digital audio editor and recording computer software application allows determination of the time-of-flight of the wave for different distances, from which the speed of sound is calculated. The result is in very good agreement with the reported value in the literature.
Advances in Audio-Based Systems to Monitor Patient Adherence and Inhaler Drug Delivery.
Taylor, Terence E; Zigel, Yaniv; De Looze, Céline; Sulaiman, Imran; Costello, Richard W; Reilly, Richard B
2018-03-01
Hundreds of millions of people worldwide have asthma and COPD. Current medications to control these chronic respiratory diseases can be administered using inhaler devices, such as the pressurized metered dose inhaler and the dry powder inhaler. Provided that they are used as prescribed, inhalers can improve patient clinical outcomes and quality of life. Poor patient inhaler adherence (both time of use and user technique) is, however, a major clinical concern and is associated with poor disease control, increased hospital admissions, and increased mortality rates, particularly in low- and middle-income countries. There are currently limited methods available to health-care professionals to objectively and remotely monitor patient inhaler adherence. This review describes recent sensor-based technologies that use audio-based approaches that show promising opportunities for monitoring inhaler adherence in clinical practice. This review discusses how one form of sensor-based technology, audio-based monitoring systems, can provide clinically pertinent information regarding patient inhaler use over the course of treatment. Audio-based monitoring can provide health-care professionals with quantitative measurements of the drug delivery of inhalers, signifying a clear clinical advantage over other methods of assessment. Furthermore, objective audio-based adherence measures can improve the predictability of patient outcomes to treatment compared with current standard methods of adherence assessment used in clinical practice. Objective feedback on patient inhaler adherence can be used to personalize treatment to the patient, which may enhance precision medicine in the treatment of chronic respiratory diseases. Copyright © 2017 American College of Chest Physicians. Published by Elsevier Inc. All rights reserved.
Quantifying auditory temporal stability in a large database of recorded music.
Ellis, Robert J; Duan, Zhiyan; Wang, Ye
2014-01-01
"Moving to the beat" is both one of the most basic and one of the most profound means by which humans (and a few other species) interact with music. Computer algorithms that detect the precise temporal location of beats (i.e., pulses of musical "energy") in recorded music have important practical applications, such as the creation of playlists with a particular tempo for rehabilitation (e.g., rhythmic gait training), exercise (e.g., jogging), or entertainment (e.g., continuous dance mixes). Although several such algorithms return simple point estimates of an audio file's temporal structure (e.g., "average tempo", "time signature"), none has sought to quantify the temporal stability of a series of detected beats. Such a method--a "Balanced Evaluation of Auditory Temporal Stability" (BEATS)--is proposed here, and is illustrated using the Million Song Dataset (a collection of audio features and music metadata for nearly one million audio files). A publically accessible web interface is also presented, which combines the thresholdable statistics of BEATS with queryable metadata terms, fostering potential avenues of research and facilitating the creation of highly personalized music playlists for clinical or recreational applications.
Improvement of information fusion-based audio steganalysis
NASA Astrophysics Data System (ADS)
Kraetzer, Christian; Dittmann, Jana
2010-01-01
In the paper we extend an existing information fusion based audio steganalysis approach by three different kinds of evaluations: The first evaluation addresses the so far neglected evaluations on sensor level fusion. Our results show that this fusion removes content dependability while being capable of achieving similar classification rates (especially for the considered global features) if compared to single classifiers on the three exemplarily tested audio data hiding algorithms. The second evaluation enhances the observations on fusion from considering only segmental features to combinations of segmental and global features, with the result of a reduction of the required computational complexity for testing by about two magnitudes while maintaining the same degree of accuracy. The third evaluation tries to build a basis for estimating the plausibility of the introduced steganalysis approach by measuring the sensibility of the models used in supervised classification of steganographic material against typical signal modification operations like de-noising or 128kBit/s MP3 encoding. Our results show that for some of the tested classifiers the probability of false alarms rises dramatically after such modifications.
Impact of audio/visual systems on pediatric sedation in magnetic resonance imaging.
Lemaire, Colette; Moran, Gerald R; Swan, Hans
2009-09-01
To evaluate the use of an audio/visual (A/V) system in pediatric patients as an alternative to sedation in magnetic resonance imaging (MRI) in terms of wait times, image quality, and patient experience. Pediatric MRI examinations from April 8 to August 11, 2008 were compared to those 1 year prior to the installation of the A/V system. Data collected included age, requisition receive date, scan date, and whether sedation was used. A posttest questionnaire was used to evaluate patient experience. Image quality was assessed by two radiologists. Over the 4 months in 2008 there was an increase of 7.2% (115; P < 0.05) of pediatric patients scanned and a decrease of 15.4%, (67; P = 0.32) requiring sedation. The average sedation wait time decreased by 33% (5.8 months) (P < 0.05). Overall, the most positively affected group was the 4-10 years. The questionnaire resulted in 84% of participants expressing a positive reaction to the A/V system. Radiological evaluation revealed no changes in image quality between A/V users and sedates. The A/V system was a successful method to reduce patient motion and obtain a quality diagnostic MRI without the use of sedation in pediatric patients. It provided a safer option, a positive experience, and decreased wait times.
Kuribayashi, Ryuma; Nittono, Hiroshi
2017-01-01
High-resolution audio has a higher sampling frequency and a greater bit depth than conventional low-resolution audio such as compact disks. The higher sampling frequency enables inaudible sound components (above 20 kHz) that are cut off in low-resolution audio to be reproduced. Previous studies of high-resolution audio have mainly focused on the effect of such high-frequency components. It is known that alpha-band power in a human electroencephalogram (EEG) is larger when the inaudible high-frequency components are present than when they are absent. Traditionally, alpha-band EEG activity has been associated with arousal level. However, no previous studies have explored whether sound sources with high-frequency components affect the arousal level of listeners. The present study examined this possibility by having 22 participants listen to two types of a 400-s musical excerpt of French Suite No. 5 by J. S. Bach (on cembalo, 24-bit quantization, 192 kHz A/D sampling), with or without inaudible high-frequency components, while performing a visual vigilance task. High-alpha (10.5-13 Hz) and low-beta (13-20 Hz) EEG powers were larger for the excerpt with high-frequency components than for the excerpt without them. Reaction times and error rates did not change during the task and were not different between the excerpts. The amplitude of the P3 component elicited by target stimuli in the vigilance task increased in the second half of the listening period for the excerpt with high-frequency components, whereas no such P3 amplitude change was observed for the other excerpt without them. The participants did not distinguish between these excerpts in terms of sound quality. Only a subjective rating of inactive pleasantness after listening was higher for the excerpt with high-frequency components than for the other excerpt. The present study shows that high-resolution audio that retains high-frequency components has an advantage over similar and indistinguishable digital sound sources in which such components are artificially cut off, suggesting that high-resolution audio with inaudible high-frequency components induces a relaxed attentional state without conscious awareness.
Learning diagnostic models using speech and language measures.
Peintner, Bart; Jarrold, William; Vergyriy, Dimitra; Richey, Colleen; Tempini, Maria Luisa Gorno; Ogar, Jennifer
2008-01-01
We describe results that show the effectiveness of machine learning in the automatic diagnosis of certain neurodegenerative diseases, several of which alter speech and language production. We analyzed audio from 9 control subjects and 30 patients diagnosed with one of three subtypes of Frontotemporal Lobar Degeneration. From this data, we extracted features of the audio signal and the words the patient used, which were obtained using our automated transcription technologies. We then automatically learned models that predict the diagnosis of the patient using these features. Our results show that learned models over these features predict diagnosis with accuracy significantly better than random. Future studies using higher quality recordings will likely improve these results.
ATS-6 - Television Relay Using Small Terminals Experiment
NASA Technical Reports Server (NTRS)
Miller, J. E.
1975-01-01
The Television Relay Using Small Terminals (TRUST) Experiment was designed to advance and promote the technology of broadcasting satellites. A constant envelope television FM signal was transmitted at C band to the ATS-6 earth coverage horn and retransmitted at 860 MHz through the 9-m antenna to a low-cost direct-readout ground station. The experiment demonstrated that high-quality television and audio can be received by low-cost direct-receive ground stations. Predetection bandwidths significantly less than predicted by Carson's rule can be utilized with minimal degradation of either monochrome or color pictures. Two separate techniques of dual audio channel transmission have been demonstrated to be suitable for low-cost applications.
Debener, Stefan; Emkes, Reiner; Volkening, Nils; Fudickar, Sebastian; Bleichner, Martin G.
2017-01-01
Objective Our aim was the development and validation of a modular signal processing and classification application enabling online electroencephalography (EEG) signal processing on off-the-shelf mobile Android devices. The software application SCALA (Signal ProCessing and CLassification on Android) supports a standardized communication interface to exchange information with external software and hardware. Approach In order to implement a closed-loop brain-computer interface (BCI) on the smartphone, we used a multiapp framework, which integrates applications for stimulus presentation, data acquisition, data processing, classification, and delivery of feedback to the user. Main Results We have implemented the open source signal processing application SCALA. We present timing test results supporting sufficient temporal precision of audio events. We also validate SCALA with a well-established auditory selective attention paradigm and report above chance level classification results for all participants. Regarding the 24-channel EEG signal quality, evaluation results confirm typical sound onset auditory evoked potentials as well as cognitive event-related potentials that differentiate between correct and incorrect task performance feedback. Significance We present a fully smartphone-operated, modular closed-loop BCI system that can be combined with different EEG amplifiers and can easily implement other paradigms. PMID:29349070
Blum, Sarah; Debener, Stefan; Emkes, Reiner; Volkening, Nils; Fudickar, Sebastian; Bleichner, Martin G
2017-01-01
Our aim was the development and validation of a modular signal processing and classification application enabling online electroencephalography (EEG) signal processing on off-the-shelf mobile Android devices. The software application SCALA (Signal ProCessing and CLassification on Android) supports a standardized communication interface to exchange information with external software and hardware. In order to implement a closed-loop brain-computer interface (BCI) on the smartphone, we used a multiapp framework, which integrates applications for stimulus presentation, data acquisition, data processing, classification, and delivery of feedback to the user. We have implemented the open source signal processing application SCALA. We present timing test results supporting sufficient temporal precision of audio events. We also validate SCALA with a well-established auditory selective attention paradigm and report above chance level classification results for all participants. Regarding the 24-channel EEG signal quality, evaluation results confirm typical sound onset auditory evoked potentials as well as cognitive event-related potentials that differentiate between correct and incorrect task performance feedback. We present a fully smartphone-operated, modular closed-loop BCI system that can be combined with different EEG amplifiers and can easily implement other paradigms.
MWAHCA: a multimedia wireless ad hoc cluster architecture.
Diaz, Juan R; Lloret, Jaime; Jimenez, Jose M; Sendra, Sandra
2014-01-01
Wireless Ad hoc networks provide a flexible and adaptable infrastructure to transport data over a great variety of environments. Recently, real-time audio and video data transmission has been increased due to the appearance of many multimedia applications. One of the major challenges is to ensure the quality of multimedia streams when they have passed through a wireless ad hoc network. It requires adapting the network architecture to the multimedia QoS requirements. In this paper we propose a new architecture to organize and manage cluster-based ad hoc networks in order to provide multimedia streams. Proposed architecture adapts the network wireless topology in order to improve the quality of audio and video transmissions. In order to achieve this goal, the architecture uses some information such as each node's capacity and the QoS parameters (bandwidth, delay, jitter, and packet loss). The architecture splits the network into clusters which are specialized in specific multimedia traffic. The real system performance study provided at the end of the paper will demonstrate the feasibility of the proposal.
Xiao, Bo; Huang, Chewei; Imel, Zac E; Atkins, David C; Georgiou, Panayiotis; Narayanan, Shrikanth S
2016-04-01
Scaling up psychotherapy services such as for addiction counseling is a critical societal need. One challenge is ensuring quality of therapy, due to the heavy cost of manual observational assessment. This work proposes a speech technology-based system to automate the assessment of therapist empathy-a key therapy quality index-from audio recordings of the psychotherapy interactions. We designed a speech processing system that includes voice activity detection and diarization modules, and an automatic speech recognizer plus a speaker role matching module to extract the therapist's language cues. We employed Maximum Entropy models, Maximum Likelihood language models, and a Lattice Rescoring method to characterize high vs. low empathic language. We estimated therapy-session level empathy codes using utterance level evidence obtained from these models. Our experiments showed that the fully automated system achieved a correlation of 0.643 between expert annotated empathy codes and machine-derived estimations, and an accuracy of 81% in classifying high vs. low empathy, in comparison to a 0.721 correlation and 86% accuracy in the oracle setting using manual transcripts. The results show that the system provides useful information that can contribute to automatic quality insurance and therapist training.
Xiao, Bo; Huang, Chewei; Imel, Zac E.; Atkins, David C.; Georgiou, Panayiotis; Narayanan, Shrikanth S.
2016-01-01
Scaling up psychotherapy services such as for addiction counseling is a critical societal need. One challenge is ensuring quality of therapy, due to the heavy cost of manual observational assessment. This work proposes a speech technology-based system to automate the assessment of therapist empathy—a key therapy quality index—from audio recordings of the psychotherapy interactions. We designed a speech processing system that includes voice activity detection and diarization modules, and an automatic speech recognizer plus a speaker role matching module to extract the therapist's language cues. We employed Maximum Entropy models, Maximum Likelihood language models, and a Lattice Rescoring method to characterize high vs. low empathic language. We estimated therapy-session level empathy codes using utterance level evidence obtained from these models. Our experiments showed that the fully automated system achieved a correlation of 0.643 between expert annotated empathy codes and machine-derived estimations, and an accuracy of 81% in classifying high vs. low empathy, in comparison to a 0.721 correlation and 86% accuracy in the oracle setting using manual transcripts. The results show that the system provides useful information that can contribute to automatic quality insurance and therapist training. PMID:28286867
NASA Astrophysics Data System (ADS)
Vassiliou, Marius S.; Sundareswaran, Venkataraman; Chen, S.; Behringer, Reinhold; Tam, Clement K.; Chan, M.; Bangayan, Phil T.; McGee, Joshua H.
2000-08-01
We describe new systems for improved integrated multimodal human-computer interaction and augmented reality for a diverse array of applications, including future advanced cockpits, tactical operations centers, and others. We have developed an integrated display system featuring: speech recognition of multiple concurrent users equipped with both standard air- coupled microphones and novel throat-coupled sensors (developed at Army Research Labs for increased noise immunity); lip reading for improving speech recognition accuracy in noisy environments, three-dimensional spatialized audio for improved display of warnings, alerts, and other information; wireless, coordinated handheld-PC control of a large display; real-time display of data and inferences from wireless integrated networked sensors with on-board signal processing and discrimination; gesture control with disambiguated point-and-speak capability; head- and eye- tracking coupled with speech recognition for 'look-and-speak' interaction; and integrated tetherless augmented reality on a wearable computer. The various interaction modalities (speech recognition, 3D audio, eyetracking, etc.) are implemented a 'modality servers' in an Internet-based client-server architecture. Each modality server encapsulates and exposes commercial and research software packages, presenting a socket network interface that is abstracted to a high-level interface, minimizing both vendor dependencies and required changes on the client side as the server's technology improves.
NASA Astrophysics Data System (ADS)
Ehmann, Andreas F.; Downie, J. Stephen
2005-09-01
The objective of the International Music Information Retrieval Systems Evaluation Laboratory (IMIRSEL) project is the creation of a large, secure corpus of audio and symbolic music data accessible to the music information retrieval (MIR) community for the testing and evaluation of various MIR techniques. As part of the IMIRSEL project, a cross-platform JAVA based visual programming environment called Music to Knowledge (M2K) is being developed for a variety of music information retrieval related tasks. The primary objective of M2K is to supply the MIR community with a toolset that provides the ability to rapidly prototype algorithms, as well as foster the sharing of techniques within the MIR community through the use of a standardized set of tools. Due to the relatively large size of audio data and the computational costs associated with some digital signal processing and machine learning techniques, M2K is also designed to support distributed computing across computing clusters. In addition, facilities to allow the integration of non-JAVA based (e.g., C/C++, MATLAB, etc.) algorithms and programs are provided within M2K. [Work supported by the Andrew W. Mellon Foundation and NSF Grants No. IIS-0340597 and No. IIS-0327371.
A Low Cost Remote Sensing System Using PC and Stereo Equipment
NASA Technical Reports Server (NTRS)
Campbell, Joel F.; Flood, Michael A.; Prasad, Narasimha S.; Hodson, Wade D.
2011-01-01
A system using a personal computer, speaker, and a microphone is used to detect objects, and make crude measurements using a carrier modulated by a pseudorandom noise (PN) code. This system can be constructed using a personal computer and audio equipment commonly found in the laboratory or at home, or more sophisticated equipment that can be purchased at reasonable cost. We demonstrate its value as an instructional tool for teaching concepts of remote sensing and digital signal processing.
Adductor spasmodic dysphonia: Relationships between acoustic indices and perceptual judgments
NASA Astrophysics Data System (ADS)
Cannito, Michael P.; Sapienza, Christine M.; Woodson, Gayle; Murry, Thomas
2003-04-01
This study investigated relationships between acoustical indices of spasmodic dysphonia and perceptual scaling judgments of voice attributes made by expert listeners. Audio-recordings of The Rainbow Passage were obtained from thirty one speakers with spasmodic dysphonia before and after a BOTOX injection of the vocal folds. Six temporal acoustic measures were obtained across 15 words excerpted from each reading sample, including both frequency of occurrence and percent time for (1) aperiodic phonation, (2) phonation breaks, and (3) fundamental frequency shifts. Visual analog scaling judgments were also obtained from six voice experts using an interactive computer interface to quantify four voice attributes (i.e., overall quality, roughness, brokenness, breathiness) in a carefully psychoacoustically controlled environment, using the same reading passages as stimuli. Number and percent aperiodicity and phonation breaks correlated significanly with perceived overall voice quality, roughness, and brokenness before and after the BOTOX injection. Breathiness was correlated with aperidocity only prior to injection, while roughness also correlated with frequency shifts following injection. Factor analysis reduced perceived attributes to two principal components: glottal squeezing and breathiness. The acoustic measures demonstrated a strong regression relationship with perceived glottal squeezing, but no regression relationship with breathiness was observed. Implications for an analysis of pathologic voices will be discussed.
The Full Monty: Locating Resources, Creating, and Presenting a Web Enhanced History Course.
ERIC Educational Resources Information Center
Bazillion, Richard J.; Braun, Connie L.
2001-01-01
Discusses how to develop a history course using the World Wide Web; course development software; full text digitized articles, electronic books, primary documents, images, and audio files; and computer equipment such as LCD projectors and interactive whiteboards. Addresses the importance of support for faculty using technology in teaching. (PAL)
Delivering Instruction via Streaming Media: A Higher Education Perspective.
ERIC Educational Resources Information Center
Mortensen, Mark; Schlieve, Paul; Young, Jon
2000-01-01
Describes streaming media, an audio/video presentation that is delivered across a network so that it is viewed while being downloaded onto the user's computer, including a continuous stream of video that can be pre-recorded or live. Discusses its use for nontraditional students in higher education and reports on implementation experiences. (LRW)
Observations on online educational materials for powder diffraction crystallography software.
DOE Office of Scientific and Technical Information (OSTI.GOV)
Toby, B. H.
2010-10-01
This article presents a series of approaches used to educate potential users of crystallographic software for powder diffraction. The approach that has been most successful in the author's opinion is the web lecture, where an audio presentation is coupled to a video-like record of the contents of the presenter's computer screen.
The Virtual Lecture Hall: Utilisation, Effectiveness and Student Perceptions
ERIC Educational Resources Information Center
Cramer, Kenneth M.; Collins, Kandice R.; Snider, Don; Fawcett, Graham
2007-01-01
We presently introduce the Virtual Lecture Hall (VLH), an instructional computer-based platform for delivering Microsoft PowerPoint slides threaded with audio clips for later review. There were 839 male and female university students enrolled in an introductory psychology class who had access to review class lectures via the VLH. This tool was…
ERIC Educational Resources Information Center
Bates, A. W.
This review of the technical possibilities of audio, television, computing, and combination media addresses the main factors influencing decisions about each technology's suitability for distance teaching, including access, costs, symbolic representation, student control, teacher control, existing structures, learning skills to be developed, and…
Code of Federal Regulations, 2011 CFR
2011-07-01
... include providing adaptive hardware and software for computers, electronic visual aids, braille devices, talking calculators, magnifiers, audio recordings and braille or large-print materials. For persons with... vision or hearing impaired, e.g., by making an announcement available in braille, in large print, or on...
Interactive Videodisc as a Component in a Multi-Method Approach to Anatomy and Physiology.
ERIC Educational Resources Information Center
Wheeler, Donald A.; Wheeler, Mary Jane
At Cuyahoga Community College (Ohio), computer-controlled interactive videodisc technology is being used as one of several instructional methods to teach anatomy and physiology. The system has the following features: audio-visual instruction, interaction with immediate feedback, self-pacing, fill-in-the-blank quizzes for testing total recall,…
ERIC Educational Resources Information Center
Parnell, Will; Bartlett, Jackie
2012-01-01
With the increased prevalence of smartphones, laptops, tablet computers, and other digital technologies, knowledge about and familiarity with the educational uses for these devices is important for early childhood teachers documenting children's learning. Teachers can use smartphones every day to take photos, record video and audio, and make…
Design and Evaluation of Simulations for the Development of Complex Decision-Making Skills.
ERIC Educational Resources Information Center
Hartley, Roger; Varley, Glen
2002-01-01
Command and Control Training Using Simulation (CACTUS) is a computer digital mapping system used by police to manage large-scale public events. Audio and video records of adaptive training scenarios using CACTUS show how the simulation develops decision-making skills for strategic and tactical event management. (SK)
ERIC Educational Resources Information Center
L'Engle, Kelly Ladin; Jackson, Christine
2008-01-01
This study examined the relationship between perceived socialization from parents, school, peers, and media and adolescents' cognitive susceptibility to initiating sexual intercourse and transition to intercourse 2 years later. Baseline and follow-up in-home Audio-Computer Assisted Self Interview surveys were completed by 854 Black and White male…
WLANs for the 21st Century Library
ERIC Educational Resources Information Center
Calamari, Cal
2009-01-01
As educational and research needs have changed, libraries have changed as well. They must meet ever-increasing demand for access to online media, subscriptions to archives, video, audio, and other content. The way a user/patron accesses this information has also changed. Gone are the days of a few hardwired desktops or computer carts. While…
ERIC Educational Resources Information Center
Yankelevich, Eleonora
2017-01-01
A variety of computing devices are available in today's classrooms, but they have not guaranteed the effective integration of technology. Nationally, teachers have ample devices, applications, productivity software, and digital audio and video tools. Despite all this, the literature suggests these tools are not employed to enhance student learning…
ERIC Educational Resources Information Center
Isakson, Carol
2006-01-01
A podcast is essentially a radio program that can be downloaded for enjoyment. Its content includes radio broadcasts, lectures, walking tours, and student-created audio projects. Most are in the standard MP3 file format that can be played on a computer, MP3 player, PDA, or newer CD or DVD players. This article presents resources for learning about…
The Power of Digital Storytelling to Support Teaching and Learning
ERIC Educational Resources Information Center
Robin, Bernard R.
2016-01-01
Although the term "digital storytelling" may not be familiar to all readers, over the last twenty years, an increasing number of educators, students and others around the world have created short movies by combining computer-based images, text, recorded audio narration, video clips and music in order to present information on various…
Pronunciation in Face-to-Face and Audio-Only Synchronous Computer-Mediated Learner Interactions
ERIC Educational Resources Information Center
Loewen, Shawn; Isbell, Daniel R.
2017-01-01
Studies of learner-learner interactions have reported varying degrees of pronunciation-focused discourse, ranging from 1% (Bowles, Toth, & Adams, 2014) to 40% (Bueno-Alastuey, 2013). Including first language (L1) background, modality, and task as variables, this study investigates the role of pronunciation in learner-learner interactions.…
ERIC Educational Resources Information Center
Rio Salado Community Coll., AZ.
Rio Salado Community College offers a variety of alternative delivery courses utilizing different forms of instructional technology (e.g., broadcast and cable television, radio, audio and video cassettes, and computer-managed instruction) for both credit and non-credit instruction. This manual provides information for student operators of a…
Methods of recording and analysing cough sounds.
Subburaj, S; Parvez, L; Rajagopalan, T G
1996-01-01
Efforts have been directed to evolve a computerized system for acquisition and multi-dimensional analysis of the cough sound. The system consists of a PC-AT486 computer with an ADC board having 12 bit resolution. The audio cough sound is acquired using a sensitive miniature microphone at a sampling rate of 8 kHz in the computer and simultaneously recorded in real time using a digital audio tape recorder which also serves as a back up. Analysis of the cough sound is done in time and frequency domains using the digitized data which provide numerical values for key parameters like cough counts, bouts, their intensity and latency. In addition, the duration of each event and cough patterns provide a unique tool which allows objective evaluation of antitussive and expectorant drugs. Both on-line and off-line checks ensure error-free performance over long periods of time. The entire system has been evaluated for sensitivity, accuracy, precision and reliability. Successful use of this system in clinical studies has established what perhaps is the first integrated approach for the objective evaluation of cough.
Key factors for a high-quality VR experience
NASA Astrophysics Data System (ADS)
Champel, Mary-Luc; Doré, Renaud; Mollet, Nicolas
2017-09-01
For many years, Virtual Reality has been presented as a promising technology that could deliver a truly new experience to users. The media and entertainment industry is now investigating the possibility to offer a video-based VR 360 experience. Nevertheless, there is a substantial risk that VR 360 could have the same fate as 3DTV if it cannot offer more than just being the next fad. The present paper aims at presenting the various quality factors required for a high-quality VR experience. More specifically, this paper will focus on the main three VR quality pillars: visual, audio and immersion.
A Planning and Development Proposal.
ERIC Educational Resources Information Center
Schachter, Rebeca
In view of the rapidly changing hardware technology along with the quality and quantity of software and general attitudes toward educational technology, the configuration of the Audio-Visual Distribution System and the Science and Engineering Library (SEL) should be flexible enough to incorporate these variables. SEL has made significant thrusts…
Analysis of DoD Usage of Multimedia Technology to Determine Requirements for Standards.
1995-03-01
1995 Accesion For NTIS CRA&I DTlC TAB Unannounced Justification By._ _ Distribution/ i D D Availability Codes Dist A-/ Avail...developing a standard architecture for their multimedia systems. When the DoD participants were asked to identify areas where standards are lack...are limited, they will sacrifice vid- eo quality in order to sustain audio quality. In order for desktop conferencing to become a market success
Movement Anticipation and EEG: Implications for BCI-Contingent Robot Therapy
Norman, Sumner L.; Dennison, Mark; Wolbrecht, Eric; Cramer, Steven C.; Srinivasan, Ramesh; Reinkensmeyer, David J.
2017-01-01
Brain-computer interfacing is a technology that has the potential to improve patient engagement in robot-assisted rehabilitation therapy. For example, movement intention reduces mu (8-13 Hz) oscillation amplitude over the sensorimotor cortex, a phenomenon referred to as event-related desynchronization (ERD). In an ERD-contingent assistance paradigm, initial BCI-enhanced robotic therapy studies have used ERD to provide robotic assistance for movement. Here we investigated how ERD changed as a function of audio-visual stimuli, overt movement from the participant, and robotic assistance. Twelve unimpaired subjects played a computer game designed for rehabilitation therapy with their fingers using the FINGER robotic exoskeleton. In the game, the participant and robot matched movement timing to audio-visual stimuli in the form of notes approaching a target on the screen set to the consistent beat of popular music. The audio-visual stimulation of the game alone did not cause ERD, before or after training. In contrast, overt movement by the subject caused ERD, whether or not the robot assisted the finger movement. Notably, ERD was also present when the subjects remained passive and the robot moved their fingers to play the game. This ERD occurred in anticipation of the passive finger movement with similar onset timing as for the overt movement conditions. These results demonstrate that ERD can be contingent on expectation of robotic assistance; that is, the brain generates an anticipatory ERD in expectation of a robot-imposed but predictable movement. This is a caveat that should be considered in designing BCIs for enhancing patient effort in roboticallyassisted therapy. PMID:26891487
Efficient techniques for wave-based sound propagation in interactive applications
NASA Astrophysics Data System (ADS)
Mehra, Ravish
Sound propagation techniques model the effect of the environment on sound waves and predict their behavior from point of emission at the source to the final point of arrival at the listener. Sound is a pressure wave produced by mechanical vibration of a surface that propagates through a medium such as air or water, and the problem of sound propagation can be formulated mathematically as a second-order partial differential equation called the wave equation. Accurate techniques based on solving the wave equation, also called the wave-based techniques, are too expensive computationally and memory-wise. Therefore, these techniques face many challenges in terms of their applicability in interactive applications including sound propagation in large environments, time-varying source and listener directivity, and high simulation cost for mid-frequencies. In this dissertation, we propose a set of efficient wave-based sound propagation techniques that solve these three challenges and enable the use of wave-based sound propagation in interactive applications. Firstly, we propose a novel equivalent source technique for interactive wave-based sound propagation in large scenes spanning hundreds of meters. It is based on the equivalent source theory used for solving radiation and scattering problems in acoustics and electromagnetics. Instead of using a volumetric or surface-based approach, this technique takes an object-centric approach to sound propagation. The proposed equivalent source technique generates realistic acoustic effects and takes orders of magnitude less runtime memory compared to prior wave-based techniques. Secondly, we present an efficient framework for handling time-varying source and listener directivity for interactive wave-based sound propagation. The source directivity is represented as a linear combination of elementary spherical harmonic sources. This spherical harmonic-based representation of source directivity can support analytical, data-driven, rotating or time-varying directivity function at runtime. Unlike previous approaches, the listener directivity approach can be used to compute spatial audio (3D audio) for a moving, rotating listener at interactive rates. Lastly, we propose an efficient GPU-based time-domain solver for the wave equation that enables wave simulation up to the mid-frequency range in tens of minutes on a desktop computer. It is demonstrated that by carefully mapping all the components of the wave simulator to match the parallel processing capabilities of the graphics processors, significant improvement in performance can be achieved compared to the CPU-based simulators, while maintaining numerical accuracy. We validate these techniques with offline numerical simulations and measured data recorded in an outdoor scene. We present results of preliminary user evaluations conducted to study the impact of these techniques on user's immersion in virtual environment. We have integrated these techniques with the Half-Life 2 game engine, Oculus Rift head-mounted display, and Xbox game controller to enable users to experience high-quality acoustics effects and spatial audio in the virtual environment.
On the Acoustics of Emotion in Audio: What Speech, Music, and Sound have in Common
Weninger, Felix; Eyben, Florian; Schuller, Björn W.; Mortillaro, Marcello; Scherer, Klaus R.
2013-01-01
Without doubt, there is emotional information in almost any kind of sound received by humans every day: be it the affective state of a person transmitted by means of speech; the emotion intended by a composer while writing a musical piece, or conveyed by a musician while performing it; or the affective state connected to an acoustic event occurring in the environment, in the soundtrack of a movie, or in a radio play. In the field of affective computing, there is currently some loosely connected research concerning either of these phenomena, but a holistic computational model of affect in sound is still lacking. In turn, for tomorrow’s pervasive technical systems, including affective companions and robots, it is expected to be highly beneficial to understand the affective dimensions of “the sound that something makes,” in order to evaluate the system’s auditory environment and its own audio output. This article aims at a first step toward a holistic computational model: starting from standard acoustic feature extraction schemes in the domains of speech, music, and sound analysis, we interpret the worth of individual features across these three domains, considering four audio databases with observer annotations in the arousal and valence dimensions. In the results, we find that by selection of appropriate descriptors, cross-domain arousal, and valence regression is feasible achieving significant correlations with the observer annotations of up to 0.78 for arousal (training on sound and testing on enacted speech) and 0.60 for valence (training on enacted speech and testing on music). The high degree of cross-domain consistency in encoding the two main dimensions of affect may be attributable to the co-evolution of speech and music from multimodal affect bursts, including the integration of nature sounds for expressive effects. PMID:23750144
Simple video format for mobile applications
NASA Astrophysics Data System (ADS)
Smith, John R.; Miao, Zhourong; Li, Chung-Sheng
2000-04-01
With the advent of pervasive computing, there is a growing demand for enabling multimedia applications on mobile devices. Large numbers of pervasive computing devices, such as personal digital assistants (PDAs), hand-held computer (HHC), smart phones, portable audio players, automotive computing devices, and wearable computers are gaining access to online information sources. However, the pervasive computing devices are often constrained along a number of dimensions, such as processing power, local storage, display size and depth, connectivity, and communication bandwidth, which makes it difficult to access rich image and video content. In this paper, we report on our initial efforts in designing a simple scalable video format with low-decoding and transcoding complexity for pervasive computing. The goal is to enable image and video access for mobile applications such as electronic catalog shopping, video conferencing, remote surveillance and video mail using pervasive computing devices.
Computer-assisted trauma care prototype.
Holzman, T G; Griffith, A; Hunter, W G; Allen, T; Simpson, R J
1995-01-01
Each year, civilian accidental injury results in 150,000 deaths and 400,000 permanent disabilities in the United States alone. The timely creation of and access to dynamically updated trauma patient information at the point of injury is critical to improving the state of care. Such information is often non-existent, incomplete, or inaccurate, resulting in less than adequate treatment by medics and the loss of precious time by medical personnel at the hospital or battalion aid station as they attempt to reassess and treat the patient. The Trauma Care Information Management System (TCIMS) is a prototype system for facilitating information flow and patient processing decisions in the difficult circumstances of civilian and military trauma care activities. The program is jointly supported by the United States Advanced Research Projects Agency (ARPA) and a consortium of universities, medical centers, and private companies. The authors' focus has been the human-computer interface for the system. We are attempting to make TCIMS powerful in the functions it delivers to its users in the field while also making it easy to understand and operate. To develop such a usable system, an approach known as user-centered design is being followed. Medical personnel themselves are collaborating with the authors in its needs analysis, design, and evaluation. Specifically, the prototype being demonstrated was designed through observation of actual civilian trauma care episodes, military trauma care exercises onboard a hospital ship, interviews with civilian and military trauma care providers, repeated evaluation of evolving prototypes by potential users, and study of the literature on trauma care and human factors engineering. This presentation at MedInfo '95 is still another avenue for soliciting guidance from medical information system experts and users. The outcome of this process is a system that provides the functions trauma care personnel desire in a manner that can be easily and accurately used in urban, rural, and military field settings. his demonstration will focus on the user interfaces for the hand-held computer device included in TCIMS, the Field Medic Associate (FMA). The FMA prototype is a ruggedized, water-resistant personal computer, weighing approximately 5 lbs. It has an LCD graphical user interface display for patient record input and output, pen-based and audio input, audio output, and wireless communications capabilities. Automatic recording and dynamic, graphical display of time-stamped trends in patient vital signs will be simulated during the demonstration. Means for accessing existing patient record information (e.g., allergies to particular medications) and updating the record with the nature of the injury, its cause, and the treatments that were administered will be shown. These will include use of an electronic pen to mark up anatoglyphs (standard drawings of human body appearing on computer screen) to show where injuries occurred and where treatments were applied, and to input textual descriptions of the nature of the injury, its cause, what treatments were administered, etc. Computer recognition of handwritten inputs will be shown. Likewise, voice annotation and audio playback of patient record information by medics and hospital personnel will be illustrated. These latter technologies free the care providers' hands to treat the patient; they can therefore provide inputs to the patient record while information is fresh in their minds. The audio playback option allows hospital personnel to select more detailed voice annotations of specific portions of the patient record by simply touching the electronic pen to a particular place where an electronic pen marking was made by a medic in the field and then listening to the medic's corresponding audio commentary. Finally, the FMA's means for assisting the medic in simultaneously managing several injured patients will be shown. (abstract truncated)
Marhefka, Stephanie L.; Santamaria, E. Karina; Leu, Cheng-Shiun; Brackis-Cott, Elizabeth; Mellins, Claude Ann
2013-01-01
Computer-assisted interview methods are increasingly popular in the assessment of sensitive behaviors (e.g., substance abuse and sexual behaviors). It has been suggested that the effect of social desirability is diminished when answering via computer, as compared to an interviewer-administered face-to-face (FTF) interview, although studies exploring this hypothesis among adolescents are rare and yield inconsistent findings. This study compared two interview modes among a sample of urban, ethnic-minority, perinatally HIV-exposed U.S. youth (baseline = 148 HIV+, 126 HIV−, ages 9–16 years; follow-up = 120 HIV+, 110 HIV−, ages 10–19 years). Participants were randomly assigned to receive a sexual behavior interview via either Audio Computer-Assisted Self-Interview (ACASI) or FTF interview. The prevalence of several sexual behaviors and participants’ reactions to the interviews were compared. Although higher rates of sexual behaviors were typically reported in the ACASI condition, the differences rarely reached statistical significance, even when limited to demographic subgroups—except for gender. Boys were significantly more likely to report several sexual behaviors in the ACASI condition compared to FTF, whereas among girls no significant differences were found between the two conditions. ACASI-assigned youth rated the interview process as easier and more enjoyable than did FTF-assigned youth, and this was fairly consistent across subgroup analyses as well. We conclude that these more positive reactions to the ACASI interview give that methodology a slight advantage, and boys may disclose more sexual behavior when using computer-assisted interviews. PMID:21604065
Dolezal, Curtis; Marhefka, Stephanie L; Santamaria, E Karina; Leu, Cheng-Shiun; Brackis-Cott, Elizabeth; Mellins, Claude Ann
2012-04-01
Computer-assisted interview methods are increasingly popular in the assessment of sensitive behaviors (e.g., substance abuse and sexual behaviors). It has been suggested that the effect of social desirability is diminished when answering via computer, as compared to an interviewer-administered face-to-face (FTF) interview, although studies exploring this hypothesis among adolescents are rare and yield inconsistent findings. This study compared two interview modes among a sample of urban, ethnic-minority, perinatally HIV-exposed U.S. youth (baseline = 148 HIV+, 126 HIV-, ages 9-16 years; follow-up = 120 HIV+, 110 HIV-, ages 10-19 years). Participants were randomly assigned to receive a sexual behavior interview via either Audio Computer-Assisted Self-Interview (ACASI) or FTF interview. The prevalence of several sexual behaviors and participants' reactions to the interviews were compared. Although higher rates of sexual behaviors were typically reported in the ACASI condition, the differences rarely reached statistical significance, even when limited to demographic subgroups--except for gender. Boys were significantly more likely to report several sexual behaviors in the ACASI condition compared to FTF, whereas among girls no significant differences were found between the two conditions. ACASI-assigned youth rated the interview process as easier and more enjoyable than did FTF-assigned youth, and this was fairly consistent across subgroup analyses as well. We conclude that these more positive reactions to the ACASI interview give that methodology a slight advantage, and boys may disclose more sexual behavior when using computer-assisted interviews.
NASA Astrophysics Data System (ADS)
Larnier, H.; Sailhac, P.; Chambodut, A.
2018-01-01
Atmospheric electromagnetic waves created by global lightning activity contain information about electrical processes of the inner and the outer Earth. Large signal-to-noise ratio events are particularly interesting because they convey information about electromagnetic properties along their path. We introduce a new methodology to automatically detect and characterize lightning-based waves using a time-frequency decomposition obtained through the application of continuous wavelet transform. We focus specifically on three types of sources, namely, atmospherics, slow tails and whistlers, that cover the frequency range 10 Hz to 10 kHz. Each wave has distinguishable characteristics in the time-frequency domain due to source shape and dispersion processes. Our methodology allows automatic detection of each type of event in the time-frequency decomposition thanks to their specific signature. Horizontal polarization attributes are also recovered in the time-frequency domain. This procedure is first applied to synthetic extremely low frequency time-series with different signal-to-noise ratios to test for robustness. We then apply it on real data: three stations of audio-magnetotelluric data acquired in Guadeloupe, oversea French territories. Most of analysed atmospherics and slow tails display linear polarization, whereas analysed whistlers are elliptically polarized. The diversity of lightning activity is finally analysed in an audio-magnetotelluric data processing framework, as used in subsurface prospecting, through estimation of the impedance response functions. We show that audio-magnetotelluric processing results depend mainly on the frequency content of electromagnetic waves observed in processed time-series, with an emphasis on the difference between morning and afternoon acquisition. Our new methodology based on the time-frequency signature of lightning-induced electromagnetic waves allows automatic detection and characterization of events in audio-magnetotelluric time-series, providing the means to assess quality of response functions obtained through processing.
McGlashan, Julian; Thuesen, Mathias Aaen; Sadolin, Cathrine
2017-05-01
We aimed to study the categorizations "Overdrive" and "Edge" from the pedagogical method Complete Vocal Technique as refiners of the often ill-defined concept of "belting" by means of audio perception, laryngostroboscopic imaging, acoustics, long-term average spectrum (LTAS), and electroglottography (EGG). This is a case-control study. Twenty singers were recorded singing sustained vowels in a "belting" quality refined by audio perception as "Overdrive" and "Edge." Two studies were performed: (1) a laryngostroboscopic examination using a videonasoendoscopic camera system (Olympus) and the Laryngostrobe program (Laryngograph); (2) a simultaneous recording of the EGG and acoustic signals using Speech Studio (Laryngograph). The images were analyzed based on consensus agreement. Statistical analysis of the acoustic, LTAS, and EGG parameters was undertaken using the Student paired t test. The two modes of singing determined by audio perception have visibly different laryngeal gestures: Edge has a more constricted setting than that of Overdrive, where the ventricular folds seem to cover more of the vocal folds, the aryepiglottic folds show a sharper edge in Edge, and the cuneiform cartilages are rolled in anteromedially. LTAS analysis shows a statistical difference, particularly after the ninth harmonic, with a coinciding first formant. The combined group showed statistical differences in shimmer, harmonics-to-noise ratio, normalized noise energy, and mean sound pressure level (P ≤ 0.05). "Belting" sounds can be categorized using audio perception into two modes of singing: "Overdrive" and "Edge." This study demonstrates consistent visibly different laryngeal gestures between these modes and with some correspondingly significant differences in LTAS, EGG, and acoustic measures. Copyright © 2017 The Voice Foundation. Published by Elsevier Inc. All rights reserved.
Description of an Audio-Based Paced Respiration Intervention for Vasomotor Symptoms
Burns, Debra S.; Drews, Michael R.; Carpenter, Janet S.
2013-01-01
Millions of women experience menopause-related hot flashes or flushes that may have a negative effect on their quality of life. Hormone therapy is an effective treatment, however, it may be contraindicated or unacceptable for some women based on previous health complications or an undesirable risk–benefit ratio. Side effects and the unacceptability of hormone therapy have created a need for behavioral interventions to reduce hot flashes. A variety of complex, multimodal behavioral, relaxation-based interventions have been studied with women (n = 88) and showed generally favorable results. However, currently extensive resource commitments reduce the translation of these interventions into standard care. Slow, deep breathing is a common component in most interventions and may be the active ingredient leading to reduced hot flashes. This article describes the content of an audio-based program designed to teach paced breathing to reduce hot flashes. Intervention content was based on skills training theory and music entrainment. The audio intervention provides an efficient way to deliver a breathing intervention that may be beneficial to other clinical populations. PMID:23914283
Telemedicine using free voice over internet protocol (VoIP) technology.
Miller, David J; Miljkovic, Nikola; Chiesa, Chad; Callahan, John B; Webb, Brad; Boedeker, Ben H
2011-01-01
Though dedicated videoteleconference (VTC) systems deliver high quality, low-latency audio and video for telemedical applications, they require expensive hardware and extensive infrastructure. The purpose of this study was to investigate free commercially available Voice over Internet Protocol (VoIP) software as a low cost alternative for telemedicine.
Using Films in the Multimedia English Class
ERIC Educational Resources Information Center
Wang, Youming
2009-01-01
With the great, constant renovation and development of various knowledge and economy, talents of compound, high quality and high skills are in urgent need in society; a new educational reform runs through the whole foreign teaching courses, including audio-visual course, speaking, reading, writing and translating courses. With the aid of…
47 CFR 101.91 - Involuntary relocation procedures.
Code of Federal Regulations, 2010 CFR
2010-10-01
... engineering, equipment, site and FCC fees, as well as any legitimate and prudent transaction expenses incurred..., reliability is measured by the percent of time the bit error rate (BER) exceeds a desired value, and for analog or digital voice transmissions, it is measured by the percent of time that audio signal quality...
Incentive Mechanisms for Peer-to-Peer Streaming
ERIC Educational Resources Information Center
Pai, Vinay
2011-01-01
The increasing popularity of high-bandwidth Internet connections has enabled new applications like the online delivery of high-quality audio and video content. Conventional server-client approaches place the entire burden of delivery on the content provider's server, making these services expensive to provide. A peer-to-peer approach allows end…
pyAudioAnalysis: An Open-Source Python Library for Audio Signal Analysis.
Giannakopoulos, Theodoros
2015-01-01
Audio information plays a rather important role in the increasing digital content that is available today, resulting in a need for methodologies that automatically analyze such content: audio event recognition for home automations and surveillance systems, speech recognition, music information retrieval, multimodal analysis (e.g. audio-visual analysis of online videos for content-based recommendation), etc. This paper presents pyAudioAnalysis, an open-source Python library that provides a wide range of audio analysis procedures including: feature extraction, classification of audio signals, supervised and unsupervised segmentation and content visualization. pyAudioAnalysis is licensed under the Apache License and is available at GitHub (https://github.com/tyiannak/pyAudioAnalysis/). Here we present the theoretical background behind the wide range of the implemented methodologies, along with evaluation metrics for some of the methods. pyAudioAnalysis has been already used in several audio analysis research applications: smart-home functionalities through audio event detection, speech emotion recognition, depression classification based on audio-visual features, music segmentation, multimodal content-based movie recommendation and health applications (e.g. monitoring eating habits). The feedback provided from all these particular audio applications has led to practical enhancement of the library.
pyAudioAnalysis: An Open-Source Python Library for Audio Signal Analysis
Giannakopoulos, Theodoros
2015-01-01
Audio information plays a rather important role in the increasing digital content that is available today, resulting in a need for methodologies that automatically analyze such content: audio event recognition for home automations and surveillance systems, speech recognition, music information retrieval, multimodal analysis (e.g. audio-visual analysis of online videos for content-based recommendation), etc. This paper presents pyAudioAnalysis, an open-source Python library that provides a wide range of audio analysis procedures including: feature extraction, classification of audio signals, supervised and unsupervised segmentation and content visualization. pyAudioAnalysis is licensed under the Apache License and is available at GitHub (https://github.com/tyiannak/pyAudioAnalysis/). Here we present the theoretical background behind the wide range of the implemented methodologies, along with evaluation metrics for some of the methods. pyAudioAnalysis has been already used in several audio analysis research applications: smart-home functionalities through audio event detection, speech emotion recognition, depression classification based on audio-visual features, music segmentation, multimodal content-based movie recommendation and health applications (e.g. monitoring eating habits). The feedback provided from all these particular audio applications has led to practical enhancement of the library. PMID:26656189
Audio-visual feedback improves the BCI performance in the navigational control of a humanoid robot
Tidoni, Emmanuele; Gergondet, Pierre; Kheddar, Abderrahmane; Aglioti, Salvatore M.
2014-01-01
Advancement in brain computer interfaces (BCI) technology allows people to actively interact in the world through surrogates. Controlling real humanoid robots using BCI as intuitively as we control our body represents a challenge for current research in robotics and neuroscience. In order to successfully interact with the environment the brain integrates multiple sensory cues to form a coherent representation of the world. Cognitive neuroscience studies demonstrate that multisensory integration may imply a gain with respect to a single modality and ultimately improve the overall sensorimotor performance. For example, reactivity to simultaneous visual and auditory stimuli may be higher than to the sum of the same stimuli delivered in isolation or in temporal sequence. Yet, knowledge about whether audio-visual integration may improve the control of a surrogate is meager. To explore this issue, we provided human footstep sounds as audio feedback to BCI users while controlling a humanoid robot. Participants were asked to steer their robot surrogate and perform a pick-and-place task through BCI-SSVEPs. We found that audio-visual synchrony between footsteps sound and actual humanoid's walk reduces the time required for steering the robot. Thus, auditory feedback congruent with the humanoid actions may improve motor decisions of the BCI's user and help in the feeling of control over it. Our results shed light on the possibility to increase robot's control through the combination of multisensory feedback to a BCI user. PMID:24987350
Influence of Immersive Human Scale Architectural Representation on Design Judgment
NASA Astrophysics Data System (ADS)
Elder, Rebecca L.
Unrealistic visual representation of architecture within our existing environments have lost all reference to the human senses. As a design tool, visual and auditory stimuli can be utilized to determine human's perception of design. This experiment renders varying building inputs within different sites, simulated with corresponding immersive visual and audio sensory cues. Introducing audio has been proven to influence the way a person perceives a space, yet most inhabitants rely strictly on their sense of vision to make design judgments. Though not as apparent, users prefer spaces that have a better quality of sound and comfort. Through a series of questions, we can begin to analyze whether a design is fit for both an acoustic and visual environment.
Home Telehealth Video Conferencing: Perceptions and Performance
Morris, Greg; Pech, Joanne; Rechter, Stuart; Carati, Colin; Kidd, Michael R
2015-01-01
Background The Flinders Telehealth in the Home trial (FTH trial), conducted in South Australia, was an action research initiative to test and evaluate the inclusion of telehealth services and broadband access technologies for palliative care patients living in the community and home-based rehabilitation services for the elderly at home. Telehealth services at home were supported by video conferencing between a therapist, nurse or doctor, and a patient using the iPad tablet. Objective The aims of this study are to identify which technical factors influence the quality of video conferencing in the home setting and to assess the impact of these factors on the clinical perceptions and acceptance of video conferencing for health care delivery into the home. Finally, we aim to identify any relationships between technical factors and clinical acceptance of this technology. Methods An action research process developed several quantitative and qualitative procedures during the FTH trial to investigate technology performance and users perceptions of the technology including measurements of signal power, data transmission throughput, objective assessment of user perceptions of videoconference quality, and questionnaires administered to clinical users. Results The effectiveness of telehealth was judged by clinicians as equivalent to or better than a home visit on 192 (71.6%, 192/268) occasions, and clinicians rated the experience of conducting a telehealth session compared with a home visit as equivalent or better in 90.3% (489/540) of the sessions. It was found that the quality of video conferencing when using a third generation mobile data service (3G) in comparison to broadband fiber-based services was concerning as 23.5% (220/936) of the calls failed during the telehealth sessions. The experimental field tests indicated that video conferencing audio and video quality was worse when using mobile data services compared with fiber to the home services. As well, statistically significant associations were found between audio/video quality and patient comfort with the technology as well as the clinician ratings for effectiveness of telehealth. Conclusions These results showed that the quality of video conferencing when using 3G-based mobile data services instead of broadband fiber-based services was less due to failed calls, audio/ video jitter, and video pixilation during the telehealth sessions. Nevertheless, clinicians felt able to deliver effective services to patients at home using 3G-based mobile data services. PMID:26381104
NASA Astrophysics Data System (ADS)
Fretz, Eric Bruce
Scaffolding is a term rooted in multiple research communities over decades of development. Customized, contingent support can be provided to learners to enable performances beyond what they can do alone. This dissertation seeks to examine how effectively scaffolds designed to promote articulation (written expressions of learner understanding) actually work, and if this effectiveness and/or the quality of the resulting models changes over time. It longitudinally examines the use of scaffolds designed into a dynamic modeling tool, as it is used by middle school science learners to create, test, and revise models of complex science phenomena like stream ecosystems. This dissertation also reviews the origins of the scaffolding construct, and summarizes conceptions of scaffolding from various lines of research. Scaffolding can be provided by both human and non-human agents, such as computers, which require specialized interface design to ensure maximum effectiveness. In the study, learners created models in four curriculum units over the seventh and eighth grade school years. Additionally, this dissertation examines the nature of the discussion learners have while using these scaffolds and the frequency and types of interpersonal scaffolds employed during the creation of models. Model quality is also examined using a rubric developed through review of prior research on assessing models and concept maps. Learner pairs' model creation sessions on a computer are captured with screen video and learner audio, and then distilled to transcripts for subsequent coding and analysis, supported by qualitative analysis software. Articulation scaffolds were found to succeed in promoting articulations and the quality of those articulations improved over time. Learner dialog associated with these written articulations is of reasonable quality but did not improve over time. Quality of model artifacts did improve over time. The overall use of scaffolding by each learner pair was contrasted with that pairs model quality, but no relationship was found. Software design and classroom implementation implications of these findings are discussed. The frequency of interpersonal scaffolding provided by teachers highlights the need to consider scaffolding holistically and synergistically, with design decisions for software tools made in light of careful analysis as to what human and non-human agents can and should each provide.
The impact of modality and working memory capacity on achievement in a multimedia environment
NASA Astrophysics Data System (ADS)
Stromfors, Charlotte M.
This study explored the impact of working memory capacity and student learning in a dual modality, multimedia environment titled Visualizing Topography. This computer-based instructional program focused on the basic skills in reading and interpreting topographic maps. Two versions of the program presented the same instructional content but varied the modality of verbal information: the audio-visual condition coordinated topographic maps and narration; the visual-visual condition provided the same topographic maps with readable text. An analysis of covariance procedure was conducted to evaluate the effects due to the two conditions in relation to working memory capacity, controlling for individual differences in spatial visualization and prior knowledge. The scores on the Figural Intersection Test were used to separate subjects into three levels in terms of their measured working memory capacity: low, medium, and high. Subjects accessed Visualizing Topography by way of the Internet and proceeded independently through the program. The program architecture was linear in format. Subjects had a minimum amount of flexibility within each of five segments, but not between segments. One hundred and fifty-one subjects were randomly assigned to either the audio-visual or the visual-visual condition. The average time spent in the program was thirty-one minutes. The results of the ANCOVA revealed a small to moderate modality effect favoring an audio-visual condition. The results also showed that subjects with low and medium working capacity benefited more from the audio-visual condition than the visual-visual condition, while subjects with a high working memory capacity did not benefit from either condition. Although splitting the data reduced group sizes, ANCOVA results by gender suggested that the audio-visual condition favored females with low working memory capacities. The results have implications for designers of educational software, the teachers who select software, and the students themselves. Splitting information into two, non-redundant sources, one audio and one visual, may effectively extend working memory capacity. This is especially significant for the student population encountering difficult science concepts that require the formation and manipulation of mental representations. It is recommended that multimedia environments be designed or selected with attention to modality conditions that facilitate student learning.
ERIC Educational Resources Information Center
Abbey, Antonia; Jacques-Tiura, Angela J.
2011-01-01
Past theory and empirical research have consistently associated a number of risk factors with sexual assault perpetration. This study extends past research by considering if the tactics which perpetrators use to obtain sex are associated with these risk factors or with characteristics of the sexual assault. Audio computer-assisted self-interviews…
ERIC Educational Resources Information Center
Beamish, Eric; And Others
This resource guide contains over 300 entries which are available through the Optimum Utilization of Resources (OUR's) exchange system. The entries describe learning materials, such as slides, video tapes, audio tapes, films, print material, and computer assisted instructional programs, which have been developed primarily by faculty of the…
Current Issues and Trends in Multidimensional Sensing Technologies for Digital Media
NASA Astrophysics Data System (ADS)
Nagata, Noriko; Ohki, Hidehiro; Kato, Kunihito; Koshimizu, Hiroyasu; Sagawa, Ryusuke; Fujiwara, Takayuki; Yamashita, Atsushi; Hashimoto, Manabu
Multidimensional sensing (MDS) technologies have numerous applications in the field of digital media, including the development of audio and visual equipment for human-computer interaction (HCI) and manufacture of data storage devices; furthermore, MDS finds applications in the fields of medicine and marketing, i.e., in e-marketing and the development of diagnosis equipment.
ERIC Educational Resources Information Center
Arnold, Savittree Rochanasmita; Padilla, Michael J.; Tunhikorn, Bupphachart
2009-01-01
In the rapidly developing digital world, technology is and will be a force in workplaces, communities, and everyday lives in the 21st century. Information and Communication Technology (ICT) including computer hardware/software, networking and other technologies such as audio, video, and other multimedia tools became learning tools for students in…
School Librarians as Technology Leaders: An Evolution in Practice
ERIC Educational Resources Information Center
Wine, Lois D.
2016-01-01
The role of school librarians has a history of radical change. School librarians adapted to take on responsibility for technology and audio-visual materials that were introduced in schools in earlier eras. With the advent of the Information Age in the middle of the 20th century and the subsequent development of personal computers and the Internet,…
1988-2000 Long-Range Plan for Technology of the Texas State Board of Education.
ERIC Educational Resources Information Center
Texas State Board of Education, Austin.
This plan plots the course for meeting educational needs in Texas through such technologies as computer-based systems, devices for storage and retrieval of massive amounts of information, telecommunications for audio, video, and information sharing, and other electronic media devised by the year 2000 that can help meet the instructional and…
Podcasts in Education: Let Their Voices Be Heard
ERIC Educational Resources Information Center
Sprague, Debra; Pixley, Cynthia
2008-01-01
One technology made possible through Web 2.0 is podcasting. Podcasts are audio, video, text, and other media files that can be played on the computer or downloaded to MP3 players. This article discusses how to create a podcast and ways to use this technology in education. Benefits and issues related to podcasting are also provided.
ERIC Educational Resources Information Center
Zalot, Alecia; Jones, Deborah J.; Kincaid, Carlye; Smith, Tasia
2009-01-01
This study replicated and extended prior research by examining neighborhood context as a moderator of the relation between the constellation of hyperactivity, impulsivity, and attention (HIA) difficulties and conduct problems among African American youth (11-16 years old; 55% girls) from single mother homes (N = 193). Using audio computer-assisted…
Extending Talk on a Prescribed Discussion Topic in a Learner-Native Speaker eTandem Learning Task
ERIC Educational Resources Information Center
Black, Emily
2017-01-01
Opportunities for language learners to access authentic input and engage in consequential interactions with native speakers of their target language abound in this era of computer mediated communication. Synchronous audio/video calling software represents one opportunity to access such input and address the challenges of developing pragmatic and…
Talk the Talk: Learner-Generated Podcasts as Catalysts for Knowledge Creation
ERIC Educational Resources Information Center
Lee, Mark J. W.; McLoughlin, Catherine; Chan, Anthony
2008-01-01
Podcasting allows audio content from one or more user-selected feeds or channels to be automatically downloaded to one's computer as it becomes available, then later transferred to a portable player for consumption at a convenient time and place. It is enjoying phenomenal growth in mainstream society, alongside other Web 2.0 technologies that…
ERIC Educational Resources Information Center
Lawless-Reljic, Sabine Karine
2010-01-01
Growing interest of educational institutions in desktop 3D graphic virtual environments for hybrid and distance education prompts questions on the efficacy of such tools. Virtual worlds, such as Second Life[R], enable computer-mediated immersion and interactions encompassing multimodal communication channels including audio, video, and text-.…
ERIC Educational Resources Information Center
Bull, Glen
2005-01-01
Podcasting allows distribution of audio files through an RSS feed. This permits users to subscribe to a series of podcasts that are automatically sent to their computer or MP3 player. The capability to receive podcasts is built into freely distributed software such as iPodder as well as the most recent version of iTunes, a free download. In this…
Use of Short Podcasts to Reinforce Learning Outcomes in Biology
ERIC Educational Resources Information Center
Aguiar, Cristina; Carvalho, Ana Amelia; Carvalho, Carla Joana
2009-01-01
Podcasts are audio or video files which can be automatically downloaded to one's computer when the episodes become available, then later transferred to a portable player for listening. The technology thereby enables the user to listen to and/or watch the content anywhere at any time. Formerly popular as radio shows, podcasting was rapidly explored…
Pathways to Drug and Sexual Risk Behaviors among Detained Adolescents
ERIC Educational Resources Information Center
Voisin, Dexter R.; Neilands, Torsten B.; Salazar, Laura F.; Crosby, Richard; DiClemente, Ralph J.
2008-01-01
This study recruited 559 youths from detention centers (mean age was 15.4 years; 50.1% of detainees were girls) to investigate pathways that link witnessing community violence in the 12 months before detainment to drug and sexual risk behaviors in the two months preceding detainment. Through the use of audio-computer-assisted technology, data were…
Navigating Turn-Taking and Conversational Repair in an Online Synchronous Course
ERIC Educational Resources Information Center
Earnshaw, Yvonne
2017-01-01
In face-to-face conversations, speaker transitions (or hand-offs) are typically seamless. In computer-mediated communication settings, speaker hand-offs can be a bit more challenging. This paper presents the results of a study of audio communication problems that occur in an online synchronous course, and how, and by whom, those problems are…
ERIC Educational Resources Information Center
Georgia Univ., Athens. Div. of Vocational Education.
A study examined teleconferencing applications that can assist educators in meeting Georgia's postsecondary vocational education needs. Three forms of teleconferencing were studied--audio conferencing, computer conferencing, and video conferencing. The study included a literature review, two pilot studies, and a survey to identify the ways in…
ERIC Educational Resources Information Center
Alexiadis, D. S.; Mitianoudis, N.
2013-01-01
Digital signal processing (DSP) has been an integral part of most electrical, electronic, and computer engineering curricula. The applications of DSP in multimedia (audio, image, video) storage, transmission, and analysis are also widely taught at both the undergraduate and post-graduate levels, as digital multimedia can be encountered in most…
Youth Self-Report of Physical and Sexual Abuse: A Latent Class Analysis
ERIC Educational Resources Information Center
Nooner, Kate B.; Litrownik, Alan J.; Thompson, Richard; Margolis, Benjamin; English, Diana J.; Knight, Elizabeth D.; Everson, Mark D.; Roesch, Scott
2010-01-01
Objective: To determine if meaningful groups of at-risk pre-adolescent youth could be identified based on their self-report of physical and sexual abuse histories. Methods: Youth participating in a consortium of ongoing longitudinal studies were interviewed using an audio-computer assisted self-interview (A-CASI) when they were approximately 12…
Here's What We Have to Say! Podcasting in the Early Childhood Classroom
ERIC Educational Resources Information Center
Berson, Ilene R.
2009-01-01
A podcast is an audio file published to the Internet for playback on mobile devices and personal computers; the meaning of the term has expanded to include video files, or "enhanced podcasts" as well. Many students are already engaged with digital technologies when they first step into early childhood classrooms. Children as young as…
ERIC Educational Resources Information Center
Jensen, Karen
2008-01-01
The library at the University of Alaska Fairbanks offers a unique equipment lending program through its Circulation Desk. The program features a wide array of equipment types, generous circulation policies, and unrestricted borrowing, enabling students, staff, and faculty to experiment with the latest in audio, video, and computer technologies,…
Federal Register 2010, 2011, 2012, 2013, 2014
2010-02-26
... public comment period and the EPA Science Advisory Board (SAB) peer review meeting, which will be... listening session is a step in EPA's revised IRIS process, announced on May 21, 2009, for developing human... audio-visual aid (e.g., laptop computer and slide projector). In general, each presentation should be no...
Federal Register 2010, 2011, 2012, 2013, 2014
2010-05-13
... science-based human health assessments to support the Agency's regulatory activities. The IRIS database... May 21, 2009, to develop human health assessments for inclusion in the IRIS database. The purpose of... if you will need audio-visual aid (e.g., lap top computer and slide projector). In general, each...
Read Naturally [R]. What Works Clearinghouse Intervention Report. Updated
ERIC Educational Resources Information Center
What Works Clearinghouse, 2013
2013-01-01
The “Read Naturally[R]” program is a supplemental reading program that aims to improve reading fluency, accuracy, and comprehension of elementary and middle school students using a combination of texts, audio CDs, and computer software. The program uses one of four products that share a common fluency-building strategy: “Read Naturally[R] Masters…
From Newton to Gates--Digital Principia
ERIC Educational Resources Information Center
Beckwith, E. George; Cunniff, Daniel T.
2008-01-01
Computers are becoming the norm for teaching and learning. The Internet gives people ready access to text, visual and audio messages from around the world. For teachers, content is critical and the future dictates the need for major changes in the role of the teacher and learner. Today's digital tools and video games have proven to be well known…
34 CFR 99.3 - What definitions apply to these regulations?
Code of Federal Regulations, 2012 CFR
2012-07-01
... (b)(3)(i) of this definition. (4) Records on a student who is 18 years of age or older, or is...)) Eligible student means a student who has reached 18 years of age or is attending an institution of..., handwriting, print, computer media, video or audio tape, film, microfilm, and microfiche. (Authority: 20 U.S.C...
34 CFR 99.3 - What definitions apply to these regulations?
Code of Federal Regulations, 2010 CFR
2010-07-01
... who is 18 years of age or older, or is attending an institution of postsecondary education, that are... by a teacher. (Authority: 20 U.S.C. 1232g(a)(4)) Eligible student means a student who has reached 18... any way, including, but not limited to, handwriting, print, computer media, video or audio tape, film...
34 CFR 99.3 - What definitions apply to these regulations?
Code of Federal Regulations, 2011 CFR
2011-07-01
... who is 18 years of age or older, or is attending an institution of postsecondary education, that are... by a teacher. (Authority: 20 U.S.C. 1232g(a)(4)) Eligible student means a student who has reached 18... any way, including, but not limited to, handwriting, print, computer media, video or audio tape, film...
34 CFR 99.3 - What definitions apply to these regulations?
Code of Federal Regulations, 2014 CFR
2014-07-01
... (b)(3)(i) of this definition. (4) Records on a student who is 18 years of age or older, or is...)) Eligible student means a student who has reached 18 years of age or is attending an institution of..., handwriting, print, computer media, video or audio tape, film, microfilm, and microfiche. (Authority: 20 U.S.C...
Federal Register 2010, 2011, 2012, 2013, 2014
2011-04-18
.... Indicate if you will need audio-visual equipment (e.g., laptop computer and slide projector). In general... diagnoses of certain tumors reported in a Ramazzini Institute methanol research study, which was cited and... (PWG) review of select studies conducted at the Institute. EPA is today releasing the draft IRIS...
ERIC Educational Resources Information Center
Bacsich, Paul D.
This paper surveys the various teleconferencing trials in the Open University, and attempts to analyze why they have been relatively unsuccessful compared to other new media initiatives such as audio cassettes and computer assisted learning. The situation is placed in the context of British and European educational use of telecommunications in…
[Volatile organic compounds (VOCs) emitted from furniture and electrical appliances].
Tanaka-Kagawa, Toshiko; Jinno, Hideto; Furukawa, Yoko; Nishimura, Tetsuji
2010-01-01
Organic chemicals are widely used as ingredients in household products. Therefore, furniture and other household products as well as building products may influence the indoor air quality. This study was performed to estimate quantitatively influence of household products on indoor air quality. Volatile organic compound (VOC) emissions were investigated for 10 products including furniture (chest, desk, dining table, sofa, cupboard) and electrical appliances (refrigerator, electric heater, desktop personal computer, liquid crystal display television and audio) by the large chamber test method (JIS A 1912) under the standard conditions of 28 degrees C, 50% relative humidity and 0.5 times/h ventilation. Emission rate of total VOC (TVOC) from the sofa showed the highest; over 7900 microg toluene-equivalent/unit/h. Relatively high TVOC emissions were observed also from desk and chest. Based on the emission rates, the impacts on the indoor TVOC were estimated by the simple model with a volume of 17.4 m3 and ventilation frequency of 0.5 times/h. The estimated TVOC increment for the sofa was 911 microg/m3, accounting for almost 230% of the provisional target value, 400 microg/m3. The values of estimated increment of toluene emitted from cupboard and styrene emitted from refrigerator were 10% and 16% of guideline values, respectively. These results revealed that VOC emissions from household products may influence significantly indoor air quality.
Quantifying Auditory Temporal Stability in a Large Database of Recorded Music
Ellis, Robert J.; Duan, Zhiyan; Wang, Ye
2014-01-01
“Moving to the beat” is both one of the most basic and one of the most profound means by which humans (and a few other species) interact with music. Computer algorithms that detect the precise temporal location of beats (i.e., pulses of musical “energy”) in recorded music have important practical applications, such as the creation of playlists with a particular tempo for rehabilitation (e.g., rhythmic gait training), exercise (e.g., jogging), or entertainment (e.g., continuous dance mixes). Although several such algorithms return simple point estimates of an audio file’s temporal structure (e.g., “average tempo”, “time signature”), none has sought to quantify the temporal stability of a series of detected beats. Such a method-a “Balanced Evaluation of Auditory Temporal Stability” (BEATS)–is proposed here, and is illustrated using the Million Song Dataset (a collection of audio features and music metadata for nearly one million audio files). A publically accessible web interface is also presented, which combines the thresholdable statistics of BEATS with queryable metadata terms, fostering potential avenues of research and facilitating the creation of highly personalized music playlists for clinical or recreational applications. PMID:25469636
Friberg, Anders; Schoonderwaldt, Erwin; Hedblad, Anton; Fabiani, Marco; Elowsson, Anders
2014-10-01
The notion of perceptual features is introduced for describing general music properties based on human perception. This is an attempt at rethinking the concept of features, aiming to approach the underlying human perception mechanisms. Instead of using concepts from music theory such as tones, pitches, and chords, a set of nine features describing overall properties of the music was selected. They were chosen from qualitative measures used in psychology studies and motivated from an ecological approach. The perceptual features were rated in two listening experiments using two different data sets. They were modeled both from symbolic and audio data using different sets of computational features. Ratings of emotional expression were predicted using the perceptual features. The results indicate that (1) at least some of the perceptual features are reliable estimates; (2) emotion ratings could be predicted by a small combination of perceptual features with an explained variance from 75% to 93% for the emotional dimensions activity and valence; (3) the perceptual features could only to a limited extent be modeled using existing audio features. Results clearly indicated that a small number of dedicated features were superior to a "brute force" model using a large number of general audio features.
A sLORETA study for gaze-independent BCI speller.
Xingwei An; Jinwen Wei; Shuang Liu; Dong Ming
2017-07-01
EEG-based BCI (brain-computer-interface) speller, especially gaze-independent BCI speller, has become a hot topic in recent years. It provides direct spelling device by non-muscular method for people with severe motor impairments and with limited gaze movement. Brain needs to conduct both stimuli-driven and stimuli-related attention in fast presented BCI paradigms for such BCI speller applications. Few researchers studied the mechanism of brain response to such fast presented BCI applications. In this study, we compared the distribution of brain activation in visual, auditory, and audio-visual combined stimuli paradigms using sLORETA (standardized low-resolution brain electromagnetic tomography). Between groups comparisons showed the importance of visual and auditory stimuli in audio-visual combined paradigm. They both contribute to the activation of brain regions, with visual stimuli being the predominate stimuli. Visual stimuli related brain region was mainly located at parietal and occipital lobe, whereas response in frontal-temporal lobes might be caused by auditory stimuli. These regions played an important role in audio-visual bimodal paradigms. These new findings are important for future study of ERP speller as well as the mechanism of fast presented stimuli.
Decoding power-spectral profiles from FMRI brain activities during naturalistic auditory experience.
Hu, Xintao; Guo, Lei; Han, Junwei; Liu, Tianming
2017-02-01
Recent studies have demonstrated a close relationship between computational acoustic features and neural brain activities, and have largely advanced our understanding of auditory information processing in the human brain. Along this line, we proposed a multidisciplinary study to examine whether power spectral density (PSD) profiles can be decoded from brain activities during naturalistic auditory experience. The study was performed on a high resolution functional magnetic resonance imaging (fMRI) dataset acquired when participants freely listened to the audio-description of the movie "Forrest Gump". Representative PSD profiles existing in the audio-movie were identified by clustering the audio samples according to their PSD descriptors. Support vector machine (SVM) classifiers were trained to differentiate the representative PSD profiles using corresponding fMRI brain activities. Based on PSD profile decoding, we explored how the neural decodability correlated to power intensity and frequency deviants. Our experimental results demonstrated that PSD profiles can be reliably decoded from brain activities. We also suggested a sigmoidal relationship between the neural decodability and power intensity deviants of PSD profiles. Our study in addition substantiates the feasibility and advantage of naturalistic paradigm for studying neural encoding of complex auditory information.
Development of a Bayesian Estimator for Audio-Visual Integration: A Neurocomputational Study
Ursino, Mauro; Crisafulli, Andrea; di Pellegrino, Giuseppe; Magosso, Elisa; Cuppini, Cristiano
2017-01-01
The brain integrates information from different sensory modalities to generate a coherent and accurate percept of external events. Several experimental studies suggest that this integration follows the principle of Bayesian estimate. However, the neural mechanisms responsible for this behavior, and its development in a multisensory environment, are still insufficiently understood. We recently presented a neural network model of audio-visual integration (Neural Computation, 2017) to investigate how a Bayesian estimator can spontaneously develop from the statistics of external stimuli. Model assumes the presence of two unimodal areas (auditory and visual) topologically organized. Neurons in each area receive an input from the external environment, computed as the inner product of the sensory-specific stimulus and the receptive field synapses, and a cross-modal input from neurons of the other modality. Based on sensory experience, synapses were trained via Hebbian potentiation and a decay term. Aim of this work is to improve the previous model, including a more realistic distribution of visual stimuli: visual stimuli have a higher spatial accuracy at the central azimuthal coordinate and a lower accuracy at the periphery. Moreover, their prior probability is higher at the center, and decreases toward the periphery. Simulations show that, after training, the receptive fields of visual and auditory neurons shrink to reproduce the accuracy of the input (both at the center and at the periphery in the visual case), thus realizing the likelihood estimate of unimodal spatial position. Moreover, the preferred positions of visual neurons contract toward the center, thus encoding the prior probability of the visual input. Finally, a prior probability of the co-occurrence of audio-visual stimuli is encoded in the cross-modal synapses. The model is able to simulate the main properties of a Bayesian estimator and to reproduce behavioral data in all conditions examined. In particular, in unisensory conditions the visual estimates exhibit a bias toward the fovea, which increases with the level of noise. In cross modal conditions, the SD of the estimates decreases when using congruent audio-visual stimuli, and a ventriloquism effect becomes evident in case of spatially disparate stimuli. Moreover, the ventriloquism decreases with the eccentricity. PMID:29046631
Older Adults Co-Creating Meaningful Individualized Social Activities Online for Healthy Ageing.
Blusi, Madeleine; Nilsson, Ingeborg; Lindgren, Helena
2018-01-01
Social isolation and loneliness among older people is a growing problem with negative effects on physical and mental health. In co-creation with older adults individualized social activities were designed where older adults through computer mediated communication were able to participate in social activities without leaving their homes. Four types of activities were designed; outdoor activity, music event, visiting a friend and leisure activity. A participatory action research design was applied, where end users together with scientists from two research fields developed, tested and evaluated online participation in the activities. Usability and safety of the systems were major concerns among older adults. The evaluation pointed out that level of simplicity, usability and audio-video quality determined the level of satisfaction with the human interaction during the activity, thereby affecting the meaningfulness of the activity. The research presented in this paper constitutes the first step in a long-term research process aiming at developing a digital coaching system that gives older adults personalized support for increasing participation in meaningful social activities.
The interplays among technology and content, immersant and VE
NASA Astrophysics Data System (ADS)
Song, Meehae; Gromala, Diane; Shaw, Chris; Barnes, Steven J.
2010-01-01
The research program aims to explore and examine the fine balance necessary for maintaining the interplays between technology and the immersant, including identifying qualities that contribute to creating and maintaining a sense of "presence" and "immersion" in an immersive virtual reality (IVR) experience. Building upon and extending previous work, we compare sitting meditation with walking meditation in a virtual environment (VE). The Virtual Meditative Walk, a new work-in-progress, integrates VR and biofeedback technologies with a self-directed, uni-directional treadmill. As immersants learn how to meditate while walking, robust, real-time biofeedback technology continuously measures breathing, skin conductance and heart rate. The physiological states of the immersant will in turn affect the audio and stereoscopic visual media through shutter glasses. We plan to test the potential benefits and limitations of this physically active form of meditation with data from a sitting form of meditation. A mixed-methods approach to testing user outcomes parallels the knowledge bases of the collaborative team: a physician, computer scientists and artists.
A Brazilian educational experiment: teleradiology on web TV.
Silva, Angélica Baptista; de Amorim, Annibal Coelho
2009-01-01
Since 2004, educational videoconferences have been held in Brazil for paediatric radiologists in training. The RUTE network has been used, a high-speed national research and education network. Twelve videoconferences were recorded by the Health Channel and transformed into TV programmes, both for conventional broadcast and for access via the Internet. Between October 2007 and December 2009 the Health Channel website registered 2378 hits. Our experience suggests that for successful recording of multipoint videoconferences, four areas are important: (1) a pre-planned script is required, for both physicians and film-makers; (2) particular care is necessary when editing the audiovisual material; (3) the audio and video equipment requires careful adjustment to preserve clinical discussions and the quality of radiology images; (4) to produce a product suitable for both TV sets and computer devices, the master tape needs to be encoded in low resolution digital video formats for Internet media (wmv and rm format for streaming, and compressed zip files for downloading) and MPEG format for DVDs.
Embedded Systems and TensorFlow Frameworks as Assistive Technology Solutions.
Mulfari, Davide; Palla, Alessandro; Fanucci, Luca
2017-01-01
In the field of deep learning, this paper presents the design of a wearable computer vision system for visually impaired users. The Assistive Technology solution exploits a powerful single board computer and smart glasses with a camera in order to allow its user to explore the objects within his surrounding environment, while it employs Google TensorFlow machine learning framework in order to real time classify the acquired stills. Therefore the proposed aid can increase the awareness of the explored environment and it interacts with its user by means of audio messages.
Breaking the news on mobile TV: user requirements of a popular mobile content
NASA Astrophysics Data System (ADS)
Knoche, Hendrik O.; Sasse, M. Angela
2006-02-01
This paper presents the results from three lab-based studies that investigated different ways of delivering Mobile TV News by measuring user responses to different encoding bitrates, image resolutions and text quality. All studies were carried out with participants watching News content on mobile devices, with a total of 216 participants rating the acceptability of the viewing experience. Study 1 compared the acceptability of a 15-second video clip at different video and audio encoding bit rates on a 3G phone at a resolution of 176x144 and an iPAQ PDA (240x180). Study 2 measured the acceptability of video quality of full feature news clips of 2.5 minutes which were recorded from broadcast TV, encoded at resolutions ranging from 120x90 to 240x180, and combined with different encoding bit rates and audio qualities presented on an iPAQ. Study 3 improved the legibility of the text included in the video simulating a separate text delivery. The acceptability of News' video quality was greatly reduced at a resolution of 120x90. The legibility of text was a decisive factor in the participants' assessment of the video quality. Resolutions of 168x126 and higher were substantially more acceptable when they were accompanied by optimized high quality text compared to proportionally scaled inline text. When accompanied by high quality text TV news clips were acceptable to the vast majority of participants at resolutions as small as 168x126 for video encoding bitrates of 160kbps and higher. Service designers and operators can apply this knowledge to design a cost-effective mobile TV experience.
Understanding and managing experiential aspects of soundscapes at Muir woods national monument.
Pilcher, Ericka J; Newman, Peter; Manning, Robert E
2009-03-01
Research has found that human-caused noise can detract from the quality of the visitor experience in national parks and related areas. Moreover, impacts to the visitor experience can be managed by formulating indicators and standards of quality as suggested in park and outdoor recreation management frameworks, such as Visitor Experience and Resource Protection (VERP), as developed by the U.S. National Park Service. The research reported in this article supports the formulation of indicators and standards of quality for human-caused noise at Muir Woods National Monument, California. Phase I identified potential indicators of quality for the soundscape of Muir Woods. A visitor "listening exercise" was conducted, where respondents identified natural and human-caused sounds heard in the park and rated the degree to which each sound was "pleasing" or "annoying." Certain visitor-caused sounds such as groups talking were heard by most respondents and were rated as annoying, suggesting that these sounds may be a good indicator of quality. Loud groups were heard by few people but were rated as highly annoying, whereas wind and water were heard by most visitors and were rated as highly pleasing. Phase II measured standards of quality for visitor-caused noise. Visitors were presented with a series of 30-second audio clips representing increasing amounts of visitor-caused sound in the park. Respondents were asked to rate the acceptability of each audio clip on a survey. Findings suggest a threshold at which visitor-caused sound is judged to be unacceptable, and is therefore considered as noise. A parallel program of sound monitoring in the park found that current levels of visitor-caused sound sometimes violate this threshold. Study findings provide an empirical basis to help formulate noise-related indicators and standards of quality in parks and related areas.
Applications in Data-Intensive Computing
DOE Office of Scientific and Technical Information (OSTI.GOV)
Shah, Anuj R.; Adkins, Joshua N.; Baxter, Douglas J.
2010-04-01
This book chapter, to be published in Advances in Computers, Volume 78, in 2010 describes applications of data intensive computing (DIC). This is an invited chapter resulting from a previous publication on DIC. This work summarizes efforts coming out of the PNNL's Data Intensive Computing Initiative. Advances in technology have empowered individuals with the ability to generate digital content with mouse clicks and voice commands. Digital pictures, emails, text messages, home videos, audio, and webpages are common examples of digital content that are generated on a regular basis. Data intensive computing facilitates human understanding of complex problems. Data-intensive applications providemore » timely and meaningful analytical results in response to exponentially growing data complexity and associated analysis requirements through the development of new classes of software, algorithms, and hardware.« less
Model-Driven Development of Interactive Multimedia Applications with MML
NASA Astrophysics Data System (ADS)
Pleuss, Andreas; Hussmann, Heinrich
There is an increasing demand for high-quality interactive applications which combine complex application logic with a sophisticated user interface, making use of individual media objects like graphics, animations, 3D graphics, audio or video. Their development is still challenging as it requires the integration of software design, user interface design, and media design.
Using Web-Conferencing with Primarily Interactive Television Courses.
ERIC Educational Resources Information Center
Collins, Mauri P.; Berge, Zane L.
Over the past seven years, Northern Arizona University (NAU) has implemented NAUnet, a professional-broadcast-quality, two-way audio, two-way video instructional television (IITV) system. The IITV system provides a face-to-face environment where students and faculty can see and hear each other and engage in discussion. Recently, several courses…
The Effect of Three Methods of Supporting the Double Bass on Muscle Tension.
ERIC Educational Resources Information Center
Dennis, Allan
1984-01-01
Using different methods of holding the double bass, college students performed Beethoven's Symphony No. 9. Audio recordings of performance were rated. Muscle tension readings from the left arm, right arm, upper back, and lower back were taken, using electromyography. Results suggest nonsignificant differences in both performance quality and muscle…
MWAHCA: A Multimedia Wireless Ad Hoc Cluster Architecture
Diaz, Juan R.; Jimenez, Jose M.; Sendra, Sandra
2014-01-01
Wireless Ad hoc networks provide a flexible and adaptable infrastructure to transport data over a great variety of environments. Recently, real-time audio and video data transmission has been increased due to the appearance of many multimedia applications. One of the major challenges is to ensure the quality of multimedia streams when they have passed through a wireless ad hoc network. It requires adapting the network architecture to the multimedia QoS requirements. In this paper we propose a new architecture to organize and manage cluster-based ad hoc networks in order to provide multimedia streams. Proposed architecture adapts the network wireless topology in order to improve the quality of audio and video transmissions. In order to achieve this goal, the architecture uses some information such as each node's capacity and the QoS parameters (bandwidth, delay, jitter, and packet loss). The architecture splits the network into clusters which are specialized in specific multimedia traffic. The real system performance study provided at the end of the paper will demonstrate the feasibility of the proposal. PMID:24737996
Archiving of Planetary Ring Data
NASA Technical Reports Server (NTRS)
Elliot, James L.
2001-01-01
Stellar occultation data provide our only Earth-based means of probing planetary rings at kilometer spatial resolution. The occultation data archive at MIT contains original data and analysis products of stellar occultations by the ring systems of the planets Jupiter, Saturn, Uranus, and Neptune observed by members of the group (and other groups) from 1977 to the present. During this time period, several media have been used to record and store the original and processed data: (1) chart records; (2) printed output, (3) audio reel tape; (4) audio cassette tape; (5) 7-track, 1/2-inch computer tape; (6) 9-track, 1/2-inch computer tape at 800, 1600, and 6250 bpi; (7) NOVA disk platters (2.5 and 5.0 Mbyte); (8) write once optical disks; (9) punched cards; and (10) read-write optical disks. With the rapid change of computer technology over this time period, some of these media have become not only obsolete, but nearly extinct. In particular, it has become nearly impossible to find any facilities that can still read 800 bpi tapes, which contain the only copies of several important data sets for the ring system of Uranus. In particular, we have an extensive ring data collection that includes data sets for the following Uranian ring occultations: U0, U11, U12, U13, U14, U25, U17, and U36.
Progress In Optical Memory Technology
NASA Astrophysics Data System (ADS)
Tsunoda, Yoshito
1987-01-01
More than 20 years have passed since the concept of optical memory was first proposed in 1966. Since then considerable progress has been made in this area together with the creation of completely new markets of optical memory in consumer and computer application areas. The first generation of optical memory was mainly developed with holographic recording technology in late 1960s and early 1970s. Considerable number of developments have been done in both analog and digital memory applications. Unfortunately, these technologies did not meet a chance to be a commercial product. The second generation of optical memory started at the beginning of 1970s with bit by bit recording technology. Read-only type optical memories such as video disks and compact audio disks have extensively investigated. Since laser diodes were first applied to optical video disk read out in 1976, there have been extensive developments of laser diode pick-ups for optical disk memory systems. The third generation of optical memory started in 1978 with bit by bit read/write technology using laser diodes. Developments of recording materials including both write-once and erasable have been actively pursued at several research institutes. These technologies are mainly focused on the optical memory systems for computer application. Such practical applications of optical memory technology has resulted in the creation of such new products as compact audio disks and computer file memories.
Patients' use of digital audio recordings in four different outpatient clinics.
Wolderslund, Maiken; Kofoed, Poul-Erik; Holst, René; Ammentorp, Jette
2015-12-01
To investigate a new technology of digital audio recording (DAR) of health consultations to provide knowledge about patients' use and evaluation of this recording method. A cross-sectional feasibility analysis of the intervention using log data from the recording platform and data from a patient-administered questionnaire. Four different outpatient clinics at a Danish hospital: Paediatrics, Orthopaedics, Internal Medicine and Urology. Two thousand seven hundred and eighty-four outpatients having their consultation audio recorded by one of 49 participating health professionals. DAR of outpatient consultations provided to patients permitting replay of their consultation either alone or together with their relatives. Replay of the consultation within 90 days from the consultation. In the adult outpatient clinics, one in every three consultations was replayed; however, the rates were significantly lower in the paediatric clinic where one in five consultations was replayed. The usage of the audio recordings was positively associated with increasing patient age and first time visits to the clinic. Patient gender influenced replays in different ways; for instance, relatives to male patients replayed recordings more often than relatives to female patients did. Approval of future recordings was high among the patients who replayed the consultation. Patients found that recording health consultations was an important information aid, and the digital recording technology was found to be feasible in routine practice. © The Author 2015. Published by Oxford University Press in association with the International Society for Quality in Health Care; all rights reserved.
Cue Integration in Categorical Tasks: Insights from Audio-Visual Speech Perception
Bejjanki, Vikranth Rao; Clayards, Meghan; Knill, David C.; Aslin, Richard N.
2011-01-01
Previous cue integration studies have examined continuous perceptual dimensions (e.g., size) and have shown that human cue integration is well described by a normative model in which cues are weighted in proportion to their sensory reliability, as estimated from single-cue performance. However, this normative model may not be applicable to categorical perceptual dimensions (e.g., phonemes). In tasks defined over categorical perceptual dimensions, optimal cue weights should depend not only on the sensory variance affecting the perception of each cue but also on the environmental variance inherent in each task-relevant category. Here, we present a computational and experimental investigation of cue integration in a categorical audio-visual (articulatory) speech perception task. Our results show that human performance during audio-visual phonemic labeling is qualitatively consistent with the behavior of a Bayes-optimal observer. Specifically, we show that the participants in our task are sensitive, on a trial-by-trial basis, to the sensory uncertainty associated with the auditory and visual cues, during phonemic categorization. In addition, we show that while sensory uncertainty is a significant factor in determining cue weights, it is not the only one and participants' performance is consistent with an optimal model in which environmental, within category variability also plays a role in determining cue weights. Furthermore, we show that in our task, the sensory variability affecting the visual modality during cue-combination is not well estimated from single-cue performance, but can be estimated from multi-cue performance. The findings and computational principles described here represent a principled first step towards characterizing the mechanisms underlying human cue integration in categorical tasks. PMID:21637344
Sweeney, Nancy M; Saarmann, Lembi; Seidman, Robert; Flagg, Joan
2006-01-01
Asynchronous online tutorials using PowerPoint slides with accompanying audio to teach practicing nurses about computers and nursing informatics were designed for this project, which awarded free continuing education units to completers. Participants had control over the advancement of slides, with the ability to repeat when desired. Graphics were kept to a minimum; thus, the program ran smoothly on computers using dial-up modems. The tutorials were marketed in live meetings and through e-mail messages on nursing listservs. Findings include that the enrollment process must be automated and instantaneous, the program must work from every type of computer and Internet connection, marketing should be live and electronic, and workshops should be offered to familiarize nurses with the online learning system.
Musical structure analysis using similarity matrix and dynamic programming
NASA Astrophysics Data System (ADS)
Shiu, Yu; Jeong, Hong; Kuo, C.-C. Jay
2005-10-01
Automatic music segmentation and structure analysis from audio waveforms based on a three-level hierarchy is examined in this research, where the three-level hierarchy includes notes, measures and parts. The pitch class profile (PCP) feature is first extracted at the note level. Then, a similarity matrix is constructed at the measure level, where a dynamic time warping (DTW) technique is used to enhance the similarity computation by taking the temporal distortion of similar audio segments into account. By processing the similarity matrix, we can obtain a coarse-grain music segmentation result. Finally, dynamic programming is applied to the coarse-grain segments so that a song can be decomposed into several major parts such as intro, verse, chorus, bridge and outro. The performance of the proposed music structure analysis system is demonstrated for pop and rock music.
Metrological digital audio reconstruction
Fadeyev,; Vitaliy, Haber [Berkeley, CA; Carl, [Berkeley, CA
2004-02-19
Audio information stored in the undulations of grooves in a medium such as a phonograph record may be reconstructed, with little or no contact, by measuring the groove shape using precision metrology methods coupled with digital image processing and numerical analysis. The effects of damage, wear, and contamination may be compensated, in many cases, through image processing and analysis methods. The speed and data handling capacity of available computing hardware make this approach practical. Two examples used a general purpose optical metrology system to study a 50 year old 78 r.p.m. phonograph record and a commercial confocal scanning probe to study a 1920's celluloid Edison cylinder. Comparisons are presented with stylus playback of the samples and with a digitally re-mastered version of an original magnetic recording. There is also a more extensive implementation of this approach, with dedicated hardware and software.
Music information retrieval in compressed audio files: a survey
NASA Astrophysics Data System (ADS)
Zampoglou, Markos; Malamos, Athanasios G.
2014-07-01
In this paper, we present an organized survey of the existing literature on music information retrieval systems in which descriptor features are extracted directly from the compressed audio files, without prior decompression to pulse-code modulation format. Avoiding the decompression step and utilizing the readily available compressed-domain information can significantly lighten the computational cost of a music information retrieval system, allowing application to large-scale music databases. We identify a number of systems relying on compressed-domain information and form a systematic classification of the features they extract, the retrieval tasks they tackle and the degree in which they achieve an actual increase in the overall speed-as well as any resulting loss in accuracy. Finally, we discuss recent developments in the field, and the potential research directions they open toward ultra-fast, scalable systems.
Exploring expressivity and emotion with artificial voice and speech technologies.
Pauletto, Sandra; Balentine, Bruce; Pidcock, Chris; Jones, Kevin; Bottaci, Leonardo; Aretoulaki, Maria; Wells, Jez; Mundy, Darren P; Balentine, James
2013-10-01
Emotion in audio-voice signals, as synthesized by text-to-speech (TTS) technologies, was investigated to formulate a theory of expression for user interface design. Emotional parameters were specified with markup tags, and the resulting audio was further modulated with post-processing techniques. Software was then developed to link a selected TTS synthesizer with an automatic speech recognition (ASR) engine, producing a chatbot that could speak and listen. Using these two artificial voice subsystems, investigators explored both artistic and psychological implications of artificial speech emotion. Goals of the investigation were interdisciplinary, with interest in musical composition, augmentative and alternative communication (AAC), commercial voice announcement applications, human-computer interaction (HCI), and artificial intelligence (AI). The work-in-progress points towards an emerging interdisciplinary ontology for artificial voices. As one study output, HCI tools are proposed for future collaboration.
Hong, OiSaeng; Eakin, Brenda L; Chin, Dal Lae; Feld, Jamie; Vogel, Stephen
2013-07-01
Noise-induced hearing loss is a significant occupational injury for firefighters exposed to intermittent noise on the job. It is important to educate firefighters about using hearing protection devices whenever they are exposed to loud noise. Computer technology is a relatively new health education approach and can be useful for tailoring specific aspects of behavioral change training. The purpose of this study is to present the development process of an Internet-based tailored intervention program and to assess its efficacy. The intervention programs were implemented for 372 firefighters (mean age = 44 years, Caucasian = 82%, male = 95%) in three states (California, Illinois, and Indiana). The efficacy was assessed from firefighters' feedback through an Internet-based survey. A multimedia Internet-based training program was developed through (a) determining program content and writing scripts, (b) developing decision-making algorithms for tailoring, (c) graphic design and audio and video productions, (d) creating computer software and a database, and (e) postproduction quality control and pilot testing. Participant feedback regarding the training has been very positive. Participants reported that they liked completing the training via computer (83%) and also that the Internet-based training program was well organized (97%), easy to use (97%), and effective (98%) and held their interest (79%). Almost all (95%) would recommend this Internet training program to other firefighters. Interactive multimedia computer technology using the Internet was a feasible mode of delivery for a hearing protection intervention among firefighters. Participants' favorable feedback strongly supports the continued utilization of this approach for designing and developing interventions to promote healthy behaviors.
Wavelet-based audio embedding and audio/video compression
NASA Astrophysics Data System (ADS)
Mendenhall, Michael J.; Claypoole, Roger L., Jr.
2001-12-01
Watermarking, traditionally used for copyright protection, is used in a new and exciting way. An efficient wavelet-based watermarking technique embeds audio information into a video signal. Several effective compression techniques are applied to compress the resulting audio/video signal in an embedded fashion. This wavelet-based compression algorithm incorporates bit-plane coding, index coding, and Huffman coding. To demonstrate the potential of this audio embedding and audio/video compression algorithm, we embed an audio signal into a video signal and then compress. Results show that overall compression rates of 15:1 can be achieved. The video signal is reconstructed with a median PSNR of nearly 33 dB. Finally, the audio signal is extracted from the compressed audio/video signal without error.
We Have Met Our Past and Our Future: Thanks for the Walk down Memory Lane
ERIC Educational Resources Information Center
Wiseman, Robert C.
2006-01-01
In this article, the author takes the readers for a walk down memory lane on the use of teaching aids. He shares his experience of the good old days of Audio Visual--opaque projector, motion pictures/films, recorders, and overhead projector. Computers have arrived, and now people can make graphics, pictures, motion pictures, and many different…
ERIC Educational Resources Information Center
Cohen, Dale; Tracy, Ryan; Cohen, Jon
2017-01-01
This study examined the effectiveness and influence on validity of a computer-based pop-up English glossary accommodation for English learners (ELs) in grades 3 and 7. In a randomized controlled trial, we administered pop-up English glossaries with audio to students taking a statewide accountability English language arts (ELA) and mathematics…
Improving Student Learning via Mobile Phone Video Content: Evidence from the BridgeIT India Project
ERIC Educational Resources Information Center
Wennersten, Matthew; Quraishy, Zubeeda Banu; Velamuri, Malathi
2015-01-01
Past efforts invested in computer-based education technology interventions have generated little evidence of affordable success at scale. This paper presents the results of a mobile phone-based intervention conducted in the Indian states of Andhra Pradesh and Tamil Nadu in 2012-13. The BridgeIT project provided a pool of audio-visual learning…
VID-R and SCAN: Tools and Methods for the Automated Analysis of Visual Records.
ERIC Educational Resources Information Center
Ekman, Paul; And Others
The VID-R (Visual Information Display and Retrieval) system that enables computer-aided analysis of visual records is composed of a film-to-television chain, two videotape recorders with complete remote control of functions, a video-disc recorder, three high-resolution television monitors, a teletype, a PDP-8, a video and audio interface, three…
ERIC Educational Resources Information Center
Manopaiboon, Chomnad; Kilmarx, Peter H.; van Griensven, Frits; Chaikummao, Supaporn; Jeeyapant, Supaporn; Limpakarnjanarat, Khanchi; Uthaiworavit, Wat
2003-01-01
Examined prevalence of and factors associated with pregnancy and abortion among vocation school students in northern Thailand. Age, current contraceptive use, early initiation of sexual intercourse, alcohol and drug use, and sexual coercion were associated with self or partner pregnancy. High rates of pregnancy and abortion indicate the need for…
ERIC Educational Resources Information Center
da Silva, André Constantino; Freire, Fernanda Maria Pereira; de Arruda, Alan Victor Pereira; da Rocha, Heloísa Vieira
2013-01-01
e-Learning environments offer content, such text, audio, video, animations, using the Web infrastructure and they are designed to users interacting with keyboard, mouse and a medium-sized screen. Mobile devices, such as smartphones and tablets, have enough computation power to render Web pages, allowing browsing the Internet and access e-Learning…
ERIC Educational Resources Information Center
Perkoski, Robert R.
2017-01-01
Computer technology provides a plethora of tools to engage students and make the classroom more interesting. Much research has been conducted on the impact of educational technology regarding instruction but little has been done on students' preferences for the type of instructor feedback (Watts, 2007). Mayer (2005) has developed an integrative,…
ERIC Educational Resources Information Center
Bequette, James W.; Brennan, Colleen
2008-01-01
Since the mid-1980s, arts policymakers in Minnesota have positioned "media arts"--defined as the "study and practice of examining human communication through photography, film or video, audio, computer or digital arts, and interactive media"--within the realm of aesthetic education and considered it one of six arts areas. This…
Project Echo: System Calculations
NASA Technical Reports Server (NTRS)
Ruthroff, Clyde L.; Jakes, William C., Jr.
1961-01-01
The primary experimental objective of Project Echo was the transmission of radio communications between points on the earth by reflection from the balloon satellite. This paper describes system calculations made in preparation for the experiment and their adaptation to the problem of interpreting the results. The calculations include path loss computations, expected audio signal-to-noise ratios, and received signal strength based on orbital parameters.
From Jefferson to Metallica to Your Campus: Copyright Issues in Student Peer-to-Peer File Sharing
ERIC Educational Resources Information Center
Cesarini, Lisa McHugh; Cesarini, Paul
2008-01-01
When Lars Ulrich, drummer for the rock group Metallica, testified before Congress about his group's lawsuit against Napster in 2000, many people who followed copyright issues in the music industry were not surprised (Ulrich, 2000). Ever since downloading audio files became as easy as clicking a few buttons on a personal computer, charges of…
ERIC Educational Resources Information Center
Cramer, Kenneth M.; Collins, Kandice R.; Snider, Don; Fawcett, Graham
2006-01-01
We further evaluated the Virtual Lecture Hall (VLH) (Cramer, Collins, Snider, & Fawcett, in press), an instructional computer-based platform to deliver PowerPoint slides threaded with audio clips for later review. Students from either an in-class or online section (ns = 810 and 74 respectively) of introductory psychology had access to live…
New Visions of Reality: Multimedia and Education.
ERIC Educational Resources Information Center
Ambron, Sueann
1986-01-01
Multimedia is a powerful tool that will change both the way we look at knowledge and our vision of reality, as well as our educational system and the business world. Multimedia as used here refers to the innovation of mixing text, audio, and video through the use of a computer. Not only will there be new products emerging from multimedia uses, but…
A quick transcribing technique for oral data
Schleicher, David
1972-01-01
Stenographic techniques offer a means for transcribing oral data accurately and efficiently. In one such application, during five Appolo lunar missions, a rough but helpful transcript was produced within minutes. Similarly, lectures, conferences, and audio tapes can be accurately transcribed as promptly as necessary. Computer programs for translating shorthand notes are being developed; they will increase both speed and accuracy of translation.
1994-01-01
with any relatively small research effort, caution must be exercised in making inferences beyond the population of specific courses taught and...Management). The adapted model is based on learning and instructionali theory. The five courses that were reconfigured in the FTP were assigned by the...distance education strategies, including audio teleconferencing, computer- based teleconferencing, and VTT. While the research is in its infancy and many
Machine-assisted editing of user-generated content
NASA Astrophysics Data System (ADS)
Cremer, Markus; Cook, Randall
2009-02-01
Over recent years user-generated content has become ubiquitously available and an attractive entertainment source for millions of end-users. Particularly for larger events, where many people use their devices to capture the action, a great number of short video clips are made available through appropriate web services. The objective of this presentation is to describe a way to combine these clips by analyzing them, and automatically reconstruct the time line in which the individual video clips were captured. This will enable people to easily create a compelling multimedia experience by leveraging multiple clips taken by different users from different angles, and across different time spans. The user will be able to shift into the role of a movie director mastering a multi-camera recording of the event. To achieve this goal, the audio portion of the video clips is analyzed, and waveform characteristics are computed with high temporal granularity in order to facilitate precise time alignment and overlap computation of the user-generated clips. Special care has to be given not only to the robustness of the selected audio features against ambient noise and various distortions, but also to the matching algorithm used to align the user-generated clips properly.
Ganesh, Attigodu Chandrashekara; Berthommier, Frédéric; Schwartz, Jean-Luc
2016-01-01
We introduce "Audio-Visual Speech Scene Analysis" (AVSSA) as an extension of the two-stage Auditory Scene Analysis model towards audiovisual scenes made of mixtures of speakers. AVSSA assumes that a coherence index between the auditory and the visual input is computed prior to audiovisual fusion, enabling to determine whether the sensory inputs should be bound together. Previous experiments on the modulation of the McGurk effect by audiovisual coherent vs. incoherent contexts presented before the McGurk target have provided experimental evidence supporting AVSSA. Indeed, incoherent contexts appear to decrease the McGurk effect, suggesting that they produce lower audiovisual coherence hence less audiovisual fusion. The present experiments extend the AVSSA paradigm by creating contexts made of competing audiovisual sources and measuring their effect on McGurk targets. The competing audiovisual sources have respectively a high and a low audiovisual coherence (that is, large vs. small audiovisual comodulations in time). The first experiment involves contexts made of two auditory sources and one video source associated to either the first or the second audio source. It appears that the McGurk effect is smaller after the context made of the visual source associated to the auditory source with less audiovisual coherence. In the second experiment with the same stimuli, the participants are asked to attend to either one or the other source. The data show that the modulation of fusion depends on the attentional focus. Altogether, these two experiments shed light on audiovisual binding, the AVSSA process and the role of attention.
Jodice, Patrick G.R.; Garman, S.L.; Collopy, Michael W.
2001-01-01
Marbled Murrelets (Brachyramphus marmoratus) are threatened seabirds that nest in coastal old-growth coniferous forests throughout much of their breeding range. Currently, observer-based audio-visual surveys are conducted at inland forest sites during the breeding season primarily to determine nesting distribution and breeding status and are being used to estimate temporal or spatial trends in murrelet detections. Our goal was to assess the feasibility of using audio-visual survey data for such monitoring. We used an intensive field-based survey effort to record daily murrelet detections at seven survey stations in the Oregon Coast Range. We then used computer-aided resampling techniques to assess the effectiveness of twelve survey strategies with varying scheduling and a sampling intensity of 4-14 surveys per breeding season to estimate known means and SDs of murrelet detections. Most survey strategies we tested failed to provide estimates of detection means and SDs that were within A?20% of actual means and SDs. Estimates of daily detections were, however, frequently estimated to within A?50% of field data with sampling efforts of 14 days/breeding season. Additional resampling analyses with statistically generated detection data indicated that the temporal variability in detection data had a great effect on the reliability of the mean and SD estimates calculated from the twelve survey strategies, while the value of the mean had little effect. Effectiveness at estimating multi-year trends in detection data was similarly poor, indicating that audio-visual surveys might be reliably used to estimate annual declines in murrelet detections of the order of 50% per year.
NASA Astrophysics Data System (ADS)
Dobre, Robert A.; Negrescu, Cristian; Stanomir, Dumitru
2016-12-01
In many situations audio recordings can decide the fate of a trial when accepted as evidence. But until they can be taken into account they must be authenticated at first, but also the quality of the targeted content (speech in most cases) must be good enough to remove any doubt. In this scope two main directions of multimedia forensics come into play: content authentication and noise reduction. This paper presents an application that is included in the latter. If someone would like to conceal their conversation, the easiest way to do it would be to turn loud the nearest audio system. In this situation, if a microphone was placed close by, the recorded signal would be apparently useless because the speech signal would be masked by the loud music signal. The paper proposes an adaptive filters based solution to remove the musical content from a previously described signal mixture in order to recover the masked vocal signal. Two adaptive filtering algorithms were tested in the proposed solution: the Normalised Least Mean Squares (NLMS) and Recursive Least Squares (RLS). Their performances in the described situation were evaluated using Simulink, compared and included in the paper.
Digital audio watermarking using moment-preserving thresholding
NASA Astrophysics Data System (ADS)
Choi, DooSeop; Jung, Hae Kyung; Choi, Hyuk; Kim, Taejeong
2007-09-01
The Moment-Preserving Thresholding technique for digital images has been used in digital image processing for decades, especially in image binarization and image compression. Its main strength lies in that the binary values that the MPT produces as a result, called representative values, are usually unaffected when the signal being thresholded goes through a signal processing operation. The two representative values in MPT together with the threshold value are obtained by solving the system of the preservation equations for the first, second, and third moment. Relying on this robustness of the representative values to various signal processing attacks considered in the watermarking context, this paper proposes a new watermarking scheme for audio signals. The watermark is embedded in the root-sum-square (RSS) of the two representative values of each signal block using the quantization technique. As a result, the RSS values are modified by scaling the signal according to the watermark bit sequence under the constraint of inaudibility relative to the human psycho-acoustic model. We also address and suggest solutions to the problem of synchronization and power scaling attacks. Experimental results show that the proposed scheme maintains high audio quality and robustness to various attacks including MP3 compression, re-sampling, jittering, and, DA/AD conversion.
Audio signal analysis for tool wear monitoring in sheet metal stamping
NASA Astrophysics Data System (ADS)
Ubhayaratne, Indivarie; Pereira, Michael P.; Xiang, Yong; Rolfe, Bernard F.
2017-02-01
Stamping tool wear can significantly degrade product quality, and hence, online tool condition monitoring is a timely need in many manufacturing industries. Even though a large amount of research has been conducted employing different sensor signals, there is still an unmet demand for a low-cost easy to set up condition monitoring system. Audio signal analysis is a simple method that has the potential to meet this demand, but has not been previously used for stamping process monitoring. Hence, this paper studies the existence and the significance of the correlation between emitted sound signals and the wear state of sheet metal stamping tools. The corrupting sources generated by the tooling of the stamping press and surrounding machinery have higher amplitudes compared to that of the sound emitted by the stamping operation itself. Therefore, a newly developed semi-blind signal extraction technique was employed as a pre-processing technique to mitigate the contribution of these corrupting sources. The spectral analysis results of the raw and extracted signals demonstrate a significant qualitative relationship between wear progression and the emitted sound signature. This study lays the basis for employing low-cost audio signal analysis in the development of a real-time industrial tool condition monitoring system.
Microphone Handling Noise: Measurements of Perceptual Threshold and Effects on Audio Quality
Kendrick, Paul; Jackson, Iain R.; Fazenda, Bruno M.; Cox, Trevor J.; Li, Francis F.
2015-01-01
A psychoacoustic experiment was carried out to test the effects of microphone handling noise on perceived audio quality. Handling noise is a problem affecting both amateurs using their smartphones and cameras, as well as professionals using separate microphones and digital recorders. The noises used for the tests were measured from a variety of devices, including smartphones, laptops and handheld microphones. The signal features that characterise these noises are analysed and presented. The sounds include various types of transient, impact noises created by tapping or knocking devices, as well as more sustained sounds caused by rubbing. During the perceptual tests, listeners auditioned speech podcasts and were asked to rate the degradation of any unwanted sounds they heard. A representative design test methodology was developed that tried to encourage everyday rather than analytical listening. Signal-to-noise ratio (SNR) of the handling noise events was shown to be the best predictor of quality degradation. Other factors such as noise type or background noise in the listening environment did not significantly affect quality ratings. Podcast, microphone type and reproduction equipment were found to be significant but only to a small extent. A model allowing the prediction of degradation from the SNR is presented. The SNR threshold at which 50% of subjects noticed handling noise was found to be 4.2 ± 0.6 dBA. The results from this work are important for the understanding of our perception of impact sound and resonant noises in recordings, and will inform the future development of an automated predictor of quality for handling noise. PMID:26473498
Quo vadimus? The 21st Century and multimedia
NASA Technical Reports Server (NTRS)
Kuhn, Allan D.
1991-01-01
The concept is related of computer driven multimedia to the NASA Scientific and Technical Information Program (STIP). Multimedia is defined here as computer integration and output of text, animation, audio, video, and graphics. Multimedia is the stage of computer based information that allows access to experience. The concepts are also drawn in of hypermedia, intermedia, interactive multimedia, hypertext, imaging, cyberspace, and virtual reality. Examples of these technology developments are given for NASA, private industry, and academia. Examples of concurrent technology developments and implementations are given to show how these technologies, along with multimedia, have put us at the threshold of the 21st century. The STI Program sees multimedia as an opportunity for revolutionizing the way STI is managed.
Computer aided systems human engineering: A hypermedia tool
NASA Technical Reports Server (NTRS)
Boff, Kenneth R.; Monk, Donald L.; Cody, William J.
1992-01-01
The Computer Aided Systems Human Engineering (CASHE) system, Version 1.0, is a multimedia ergonomics database on CD-ROM for the Apple Macintosh II computer, being developed for use by human system designers, educators, and researchers. It will initially be available on CD-ROM and will allow users to access ergonomics data and models stored electronically as text, graphics, and audio. The CASHE CD-ROM, Version 1.0 will contain the Boff and Lincoln (1988) Engineering Data Compendium, MIL-STD-1472D and a unique, interactive simulation capability, the Perception and Performance Prototyper. Its features also include a specialized data retrieval, scaling, and analysis capability and the state of the art in information retrieval, browsing, and navigation.
Improving student learning via mobile phone video content: Evidence from the BridgeIT India project
NASA Astrophysics Data System (ADS)
Wennersten, Matthew; Quraishy, Zubeeda Banu; Velamuri, Malathi
2015-08-01
Past efforts invested in computer-based education technology interventions have generated little evidence of affordable success at scale. This paper presents the results of a mobile phone-based intervention conducted in the Indian states of Andhra Pradesh and Tamil Nadu in 2012-13. The BridgeIT project provided a pool of audio-visual learning materials organised in accordance with a system of syllabi pacing charts. Teachers of Standard 5 and 6 English and Science classes were notified of the availability of new videos via text messages (SMS), which they downloaded onto their phones using an open-source application and showed, with suggested activities, to students on a TV screen using a TV-out cable. In their evaluation of this project, the authors of this paper found that the test scores of children who experienced the intervention improved by 0.36 standard deviations in English and 0.98 standard deviations in Science in Andhra Pradesh, relative to students in similar classrooms who did not experience the intervention. Differences between treatment and control schools in Tamil Nadu were less marked. The intervention was also cost-effective, relative to other computer-based interventions. Based on these results, the authors argue that is possible to use mobile phones to produce a strong positive and statistically significant effect in terms of teaching and learning quality across a large number of classrooms in India at a lower cost per student than past computer-based interventions.
Ad Hoc Selection of Voice over Internet Streams
NASA Technical Reports Server (NTRS)
Macha, Mitchell G. (Inventor); Bullock, John T. (Inventor)
2014-01-01
A method and apparatus for a communication system technique involving ad hoc selection of at least two audio streams is provided. Each of the at least two audio streams is a packetized version of an audio source. A data connection exists between a server and a client where a transport protocol actively propagates the at least two audio streams from the server to the client. Furthermore, software instructions executable on the client indicate a presence of the at least two audio streams, allow selection of at least one of the at least two audio streams, and direct the selected at least one of the at least two audio streams for audio playback.
Ad Hoc Selection of Voice over Internet Streams
NASA Technical Reports Server (NTRS)
Macha, Mitchell G. (Inventor); Bullock, John T. (Inventor)
2008-01-01
A method and apparatus for a communication system technique involving ad hoc selection of at least two audio streams is provided. Each of the at least two audio streams is a packetized version of an audio source. A data connection exists between a server and a client where a transport protocol actively propagates the at least two audio streams from the server to the client. Furthermore, software instructions executable on the client indicate a presence of the at least two audio streams, allow selection of at least one of the at least two audio streams, and direct the selected at least one of the at least two audio streams for audio playback.
Theory and Practice: How Filming "Learning in the Real World" Helps Students Make the Connection
ERIC Educational Resources Information Center
Commander, Nannette Evans; Ward, Teresa E.; Zabrucky, Karen M.
2012-01-01
This article describes an assignment, titled "Learning in the Real World," designed for graduate students in a learning theory course. Students work in small groups to create high quality audio-visual films that present "real learning" through interviews and/or observations of learners. Students select topics relevant to theories we are discussing…
Audio direct broadcast satellites
NASA Technical Reports Server (NTRS)
Miller, J. E.
1983-01-01
Satellite sound broadcasting is, as the name implies, the use of satellite techniques and technology to broadcast directly from space to low-cost, consumer-quality receivers the types of sound programs commonly received in the AM and FM broadcast bands. It would be a ubiquitous service available to the general public in the home, in the car, and out in the open.
ERIC Educational Resources Information Center
Alpiste Penalba, Francisco; Rojas-Rajs, Teresa; Lorente, Pedro; Iglesias, Francisco; Fernández, Joaquín; Monguet, Josep
2013-01-01
The Opera eLearning project developed a solution for opera singing distance lessons at the graduate level, using high bandwidth to deliver a quality audio and video experience that has been evaluated by singing teachers, chorus and orchestra directors, singers and other professional musicians. Prior to finding a technological model that suits the…
ERIC Educational Resources Information Center
Karim, Salma; Kama, Mustafa Azad; Islam, Md. Mayenul
2001-01-01
Examines the role and effectiveness of radio and television programs in distance and open learning systems based on a survey of students at Bangladesh Open University. Offers recommendations for improvements, including extending the time span, ensuring the quality of presentations, teacher involvement, and audio and video cassette…
Performance of AESP Transmission/Reception Equipment; Summer and Fall, 1974 Technical Report No. 5.
ERIC Educational Resources Information Center
Bramble, William J.; And Others
Appalachian Education Satellite Project (AESP), using the ATS-6 satellite, has designed a variety of multi-media learning activities intended to upgrade the quality of instruction in Appalachia. Four modes of communication (televised programs, four-channel audio review, four-channel data collection and analysis, and VHF-teletype relay system) were…
Audio in Courseware: Design Knowledge Issues.
ERIC Educational Resources Information Center
Aarntzen, Diana
1993-01-01
Considers issues that need to be addressed when incorporating audio in courseware design. Topics discussed include functions of audio in courseware; the relationship between auditive and visual information; learner characteristics in relation to audio; events of instruction; and audio characteristics, including interactivity and speech technology.…
A Virtual Audio Guidance and Alert System for Commercial Aircraft Operations
NASA Technical Reports Server (NTRS)
Begault, Durand R.; Wenzel, Elizabeth M.; Shrum, Richard; Miller, Joel; Null, Cynthia H. (Technical Monitor)
1996-01-01
Our work in virtual reality systems at NASA Ames Research Center includes the area of aurally-guided visual search, using specially-designed audio cues and spatial audio processing (also known as virtual or "3-D audio") techniques (Begault, 1994). Previous studies at Ames had revealed that use of 3-D audio for Traffic Collision Avoidance System (TCAS) advisories significantly reduced head-down time, compared to a head-down map display (0.5 sec advantage) or no display at all (2.2 sec advantage) (Begault, 1993, 1995; Begault & Pittman, 1994; see Wenzel, 1994, for an audio demo). Since the crew must keep their head up and looking out the window as much as possible when taxiing under low-visibility conditions, and the potential for "blunder" is increased under such conditions, it was sensible to evaluate the audio spatial cueing for a prototype audio ground collision avoidance warning (GCAW) system, and a 3-D audio guidance system. Results were favorable for GCAW, but not for the audio guidance system.
The priming function of in-car audio instruction.
Keyes, Helen; Whitmore, Antony; Naneva, Stanislava; McDermott, Daragh
2018-05-01
Studies to date have focused on the priming power of visual road signs, but not the priming potential of audio road scene instruction. Here, the relative priming power of visual, audio, and multisensory road scene instructions was assessed. In a lab-based study, participants responded to target road scene turns following visual, audio, or multisensory road turn primes which were congruent or incongruent to the primes in direction, or control primes. All types of instruction (visual, audio, and multisensory) were successful in priming responses to a road scene. Responses to multisensory-primed targets (both audio and visual) were faster than responses to either audio or visual primes alone. Incongruent audio primes did not affect performance negatively in the manner of incongruent visual or multisensory primes. Results suggest that audio instructions have the potential to prime drivers to respond quickly and safely to their road environment. Peak performance will be observed if audio and visual road instruction primes can be timed to co-occur.
Audio-visual interactions in environment assessment.
Preis, Anna; Kociński, Jędrzej; Hafke-Dys, Honorata; Wrzosek, Małgorzata
2015-08-01
The aim of the study was to examine how visual and audio information influences audio-visual environment assessment. Original audio-visual recordings were made at seven different places in the city of Poznań. Participants of the psychophysical experiments were asked to rate, on a numerical standardized scale, the degree of comfort they would feel if they were in such an environment. The assessments of audio-visual comfort were carried out in a laboratory in four different conditions: (a) audio samples only, (b) original audio-visual samples, (c) video samples only, and (d) mixed audio-visual samples. The general results of this experiment showed a significant difference between the investigated conditions, but not for all the investigated samples. There was a significant improvement in comfort assessment when visual information was added (in only three out of 7 cases), when conditions (a) and (b) were compared. On the other hand, the results show that the comfort assessment of audio-visual samples could be changed by manipulating the audio rather than the video part of the audio-visual sample. Finally, it seems, that people could differentiate audio-visual representations of a given place in the environment based rather of on the sound sources' compositions than on the sound level. Object identification is responsible for both landscape and soundscape grouping. Copyright © 2015. Published by Elsevier B.V.
47 CFR 73.403 - Digital audio broadcasting service requirements.
Code of Federal Regulations, 2012 CFR
2012-10-01
... 47 Telecommunication 4 2012-10-01 2012-10-01 false Digital audio broadcasting service requirements... SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast radio stations using IBOC must transmit at least one over-the-air digital audio...
47 CFR 73.403 - Digital audio broadcasting service requirements.
Code of Federal Regulations, 2011 CFR
2011-10-01
... 47 Telecommunication 4 2011-10-01 2011-10-01 false Digital audio broadcasting service requirements... SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast radio stations using IBOC must transmit at least one over-the-air digital audio...
47 CFR 73.403 - Digital audio broadcasting service requirements.
Code of Federal Regulations, 2014 CFR
2014-10-01
... 47 Telecommunication 4 2014-10-01 2014-10-01 false Digital audio broadcasting service requirements... SERVICES RADIO BROADCAST SERVICES Digital Audio Broadcasting § 73.403 Digital audio broadcasting service requirements. (a) Broadcast radio stations using IBOC must transmit at least one over-the-air digital audio...