Science.gov

Sample records for acoustic vowel space

  1. Reference Data for the American English Acoustic Vowel Space

    ERIC Educational Resources Information Center

    Flipsen, Peter, Jr.; Lee, Sungbok

    2012-01-01

    Reference data for the acoustic vowel space area (VSA) in children and adolescents do not currently appear to be available in a form suitable for normative comparisons. In the current study, individual speaker formant data for the four corner vowels of American English (/i, u, ae, [alpha]/) were used to compute individual speaker VSAs. The sample…

  2. Vowel Acoustic Space Development in Children: A Synthesis of Acoustic and Anatomic Data

    ERIC Educational Resources Information Center

    Vorperian, Houri K.; Kent, Ray D.

    2007-01-01

    Purpose: This article integrates published acoustic data on the development of vowel production. Age specific data on formant frequencies are considered in the light of information on the development of the vocal tract (VT) to create an anatomic-acoustic description of the maturation of the vowel acoustic space for English. Method: Literature…

  3. Vowel Acoustic Space Development in Children: A Synthesis of Acoustic and Anatomic Data

    PubMed Central

    Vorperian, Houri K.; Kent, Ray D.

    2008-01-01

    Purpose This article integrates published acoustic data on the development of vowel production. Age specific data on formant-frequencies are considered in the light of information on the development of the vocal tract (VT) to create an anatomic-acoustic description of the maturation of the vowel acoustic space for English. Method Literature searches identified 14 studies reporting data on vowel formant-frequencies. Data on corner vowels are summarized graphically to show age/sex related changes in the area and shape of the traditional vowel quadrilateral. Conclusions Vowel development is expressed as: (a) establishment of a language-appropriate acoustic representation (e.g., F1-F2 quadrilateral or F1-F2-F3 space), (b) gradual reduction in formant-frequencies and F1-F2 area with age, (c) reduction in formant-frequency variability, (d) emergence of male-female differences in formant-frequency by age 4 years with more apparent differences by 8 years, (e) jumps in formant-frequency at ages corresponding to growth spurts of the VT, and (f) a decline of f0 after age 1, with the decline being more rapid during early childhood and adolescence. Questions remain about optimal procedures for VT normalization, and the exact relationship between VT growth and formant-frequencies. Comments are included on nasalization and vocal fundamental-frequency as they relate to the development of vowel production. PMID:18055771

  4. The influence of phonetic context and formant measurement location on acoustic vowel space

    NASA Astrophysics Data System (ADS)

    Turner, Greg S.; Hutchings, David T.; Sylvester, Betsy; Weismer, Gary

    2003-04-01

    One way of depicting vowel production is by describing vowels within an F1/F2 acoustic vowel space. This acoustic measure illustrates the dispersion of F1 and F2 values at a specific moment in time (e.g., the temporal midpoint of a vowel) for the vowels of a given language. This measure has recently been used to portray vowel production in individuals with communication disorders such as dysarthria and is moderately related to the severity of the speech disorder. Studies aimed at identifying influential factors effecting measurement stability of vowel space have yet to be completed. The focus of the present study is to evaluate the influence of phonetic context and spectral measurement location on vowel space in a group of neurologically normal American English speakers. For this study, vowel space was defined in terms of the dispersion of the four corner vowels produced within a CVC syllable frame, where C includes six stop consonants in all possible combinations with each vowel. Spectral measures were made at the midpoint and formant extremes of the vowels. A discussion will focus on individual and group variation in vowel space as a function of phonetic context and temporal measurement location.

  5. A modeling investigation of vowel-to-vowel movement planning in acoustic and muscle spaces

    NASA Astrophysics Data System (ADS)

    Zandipour, Majid

    The primary objective of this research was to explore the coordinate space in which speech movements are planned. A two dimensional biomechanical model of the vocal tract (tongue, lips, jaw, and pharynx) was constructed based on anatomical and physiological data from a subject. The model transforms neural command signals into the actions of muscles. The tongue was modeled by a 221-node finite element mesh. Each of the eight tongue muscles defined within the mesh was controlled by a virtual muscle model. The other vocal-tract components were modeled as simple 2nd-order systems. The model's geometry was adapted to a speaker, using MRI scans of the speaker's vocal tract. The vocal tract model, combined with an adaptive controller that consisted of a forward model (mapping 12-dimensional motor commands to a 64-dimensional acoustic spectrum) and an inverse model (mapping acoustic trajectories to motor command trajectories), was used to simulate and explore the implications of two planning hypotheses: planning in motor space vs. acoustic space. The acoustic, kinematic, and muscle activation (EMG) patterns of vowel-to-vowel sequences generated by the model were compared to data from the speaker whose acoustic, kinematic and EMG were also recorded. The simulation results showed that: (a) modulations of the motor commands effectively accounted for the effects of speaking rate on EMG, kinematic, and acoustic outputs; (b) the movement and acoustic trajectories were influenced by vocal tract biomechanics; and (c) both planning schemes produced similar articulatory movement, EMG, muscle length, force, and acoustic trajectories, which were also comparable to the subject's data under normal speaking conditions. In addition, the effects of a bite-block on measured EMG, kinematics and formants were simulated by the model. Acoustic planning produced successful simulations but motor planning did not. The simulation results suggest that with somatosensory feedback but no auditory

  6. Subglottal coupling and vowel space

    NASA Astrophysics Data System (ADS)

    Chi, Xuemin; Sonderegger, Morgan

    2001-05-01

    A model of acoustic coupling between the oral and subglottal cavities predicts discontinuities in vowel formant prominences near resonances of the subglottal system. One discontinuity occurs near 1300-1500 Hz, suggesting the hypothesis that this is a quantal effect [K. N. Stevens, J. Phonetics 17, 3-46 (1989)] dividing speakers' front and back vowels. Recordings of English vowels (in /hVd/ environments) for several male and female speakers were made, while an accelerometer attached to the neck area was used to capture the subglottal waveform. Statistics on our subglottal resonance measurements are given and compared with prior work. Qualitative agreement is shown between the resonator model and diphthong data with time-varying F2 for several speakers. Comparison of the second vowel formant and second subglottal formant tracks across all speakers, analysis of the formant spaces spanned by each speaker's vowel data, and a survey of vowel formant data for a sample of the world's languages support the possibility that a speaker's second subglottal resonance divides front and back vowels. Possible implications for theories of vowel inventory structure [e.g., J. Lijencrants and B. Lindblom, Language 48, 839-862 (1972)] are discussed. [Work supported by NIH Grant DC00075.

  7. Articulation Rate and Vowel Space Characteristics of Young Males with Fragile X Syndrome: Preliminary Acoustic Findings

    ERIC Educational Resources Information Center

    Zajac, David J.; Roberts, Joanne E.; Hennon, Elizabeth A.; Harris, Adrianne A.; Barnes, Elizabeth F.; Misenheimer, Jan

    2006-01-01

    Purpose: Increased speaking rate is a commonly reported perceptual characteristic among males with fragile X syndrome (FXS). The objective of this preliminary study was to determine articulation rate--one component of perceived speaking rate--and vowel space characteristics of young males with FXS. Method: Young males with FXS (n = 38), …

  8. Acoustic characteristics of vowel sounds in patients with Parkinson disease.

    PubMed

    Bang, Young-Im; Min, Kyunghoon; Sohn, Young H; Cho, Sung-Rae

    2013-01-01

    The purpose of this study was to define the acoustic voice and speech characteristics of patients with Parkinson disease (PD). Seven female patients with PD and seven female healthy controls participated in this study. Each subject was instructed to vocalize extended corner vowels (/a/, /e/, /i/, /u/) three times for at least 5 seconds at a comfortable voice loudness and tone. The voice was analyzed using the Praat program. As a result, female patients with PD showed a significant increase in jitter and noise-to-harmonics ratio (NHR). In addition, F1 and F2 among the PD patients demonstrated asymmetric centralization of unrounded vowels (/a/, /e/, /i/) in high/low/front/back positions of the tongue, consequently leading to a significant decrease in vowel space area, compared to healthy controls. This study showed the acoustic characteristics of vowel sounds not only by laryngeal variables such as abnormal jitter and NHR, but also by articulatory variables such as asymmetric centralization and reduced vowel space area in female patients with PD. Therefore, it is important to use these objective and sensitive variables to evaluate the status or severity of hypokinetic dysarthria in patients with PD.

  9. Contextual variation in the acoustic and perceptual similarity of North German and American English vowels

    NASA Astrophysics Data System (ADS)

    Strange, Winifred; Bohn, Ocke-Schwen; Nishi, Kanae; Trent, Sonja A.

    2005-09-01

    Strange et al. [J. Acoust. Soc. Am. 115, 1791-1807 (2004)] reported that North German (NG) front-rounded vowels in hVp syllables were acoustically intermediate between front and back American English (AE) vowels. However, AE listeners perceptually assimilated them as poor exemplars of back AE vowels. In this study, speaker- and context-independent cross-language discriminant analyses of NG and AE vowels produced in CVC syllables (C=labial, alveolar, velar stops) in sentences showed that NG front-rounded vowels fell within AE back-vowel distributions, due to the ``fronting'' of AE back vowels in alveolar/velar contexts. NG [smcapi, e, ɛ, openo] were located relatively ``higher'' in acoustic vowel space than their AE counterparts and varied in cross-language similarity across consonantal contexts. In a perceptual assimilation task, naive listeners classified NG vowels in terms of native AE categories and rated their goodness on a 7-point scale (very foreign to very English sounding). Both front- and back-rounded NG vowels were perceptually assimilated overwhelmingly to back AE categories and judged equally good exemplars. Perceptual assimilation patterns did not vary with context, and were not always predictable from acoustic similarity. These findings suggest that listeners adopt a context-independent strategy when judging the cross-language similarity of vowels produced and presented in continuous speech contexts.

  10. Temporal and acoustic characteristics of Greek vowels produced by adults with cerebral palsy

    NASA Astrophysics Data System (ADS)

    Botinis, Antonis; Orfanidou, Ioanna; Fourakis, Marios; Fourakis, Marios

    2005-09-01

    The present investigation examined the temporal and spectral characteristics of Greek vowels as produced by speakers with intact (NO) versus cerebral palsy affected (CP) neuromuscular systems. Six NO and six CP native speakers of Greek produced the Greek vowels [i, e, a, o, u] in the first syllable of CVCV nonsense words in a short carrier phrase. Stress could be on either the first or second syllable. There were three female and three male speakers in each group. In terms of temporal characteristics, the results showed that: vowels produced by CP speakers were longer than vowels produced by NO speakers; stressed vowels were longer than unstressed vowels; vowels produced by female speakers were longer than vowels produced by male speakers. In terms of spectral characteristics the results showed that the vowel space of the CP speakers was smaller than that of the NO speakers. This is similar to the results recently reported by Liu et al. [J. Acoust. Soc. Am. 117, 3879-3889 (2005)] for CP speakers of Mandarin. There was also a reduction of the acoustic vowel space defined by unstressed vowels, but this reduction was much more pronounced in the vowel productions of CP speakers than NO speakers.

  11. Acoustic properties of vowel production in prelingually deafened Mandarin-speaking children with cochlear implants.

    PubMed

    Yang, Jing; Brown, Emily; Fox, Robert A; Xu, Li

    2015-11-01

    The present study examined the acoustic features of vowel production in Mandarin-speaking children with cochlear implants (CIs). The subjects included 14 native Mandarin-speaking, prelingually deafened children with CIs (2.9-8.3 yr old) and 60 age-matched, normal-hearing (NH) children (3.1-9.0 years old). Each subject produced a list of monosyllables containing seven Mandarin vowels: [i, a, u, y, ɤ, ʅ, ɿ]. Midpoint F1 and F2 of each vowel token were extracted and normalized to eliminate the effects of different vocal tract sizes. Results showed that the CI children produced significantly longer vowels and less compact vowel categories than the NH children did. The CI children's acoustic vowel space was reduced due to a retracted production of the vowel [i]. The vowel space area showed a strong negative correlation with age at implantation (r = -0.80). The analysis of acoustic distance showed that the CI children produced corner vowels [a, u] similarly to the NH children, but other vowels (e.g., [ʅ, ɿ]) differently from the NH children, which suggests that CI children generally follow a similar developmental path of vowel acquisition as NH children. These findings highlight the importance of early implantation and have implications in clinical aural habilitation in young children with CIs.

  12. Acoustic properties of vowel production in prelingually deafened Mandarin-speaking children with cochlear implants.

    PubMed

    Yang, Jing; Brown, Emily; Fox, Robert A; Xu, Li

    2015-11-01

    The present study examined the acoustic features of vowel production in Mandarin-speaking children with cochlear implants (CIs). The subjects included 14 native Mandarin-speaking, prelingually deafened children with CIs (2.9-8.3 yr old) and 60 age-matched, normal-hearing (NH) children (3.1-9.0 years old). Each subject produced a list of monosyllables containing seven Mandarin vowels: [i, a, u, y, ɤ, ʅ, ɿ]. Midpoint F1 and F2 of each vowel token were extracted and normalized to eliminate the effects of different vocal tract sizes. Results showed that the CI children produced significantly longer vowels and less compact vowel categories than the NH children did. The CI children's acoustic vowel space was reduced due to a retracted production of the vowel [i]. The vowel space area showed a strong negative correlation with age at implantation (r = -0.80). The analysis of acoustic distance showed that the CI children produced corner vowels [a, u] similarly to the NH children, but other vowels (e.g., [ʅ, ɿ]) differently from the NH children, which suggests that CI children generally follow a similar developmental path of vowel acquisition as NH children. These findings highlight the importance of early implantation and have implications in clinical aural habilitation in young children with CIs. PMID:26627755

  13. Degraded Vowel Acoustics and the Perceptual Consequences in Dysarthria

    NASA Astrophysics Data System (ADS)

    Lansford, Kaitlin L.

    Distorted vowel production is a hallmark characteristic of dysarthric speech, irrespective of the underlying neurological condition or dysarthria diagnosis. A variety of acoustic metrics have been used to study the nature of vowel production deficits in dysarthria; however, not all demonstrate sensitivity to the exhibited deficits. Less attention has been paid to quantifying the vowel production deficits associated with the specific dysarthrias. Attempts to characterize the relationship between naturally degraded vowel production in dysarthria with overall intelligibility have met with mixed results, leading some to question the nature of this relationship. It has been suggested that aberrant vowel acoustics may be an index of overall severity of the impairment and not an "integral component" of the intelligibility deficit. A limitation of previous work detailing perceptual consequences of disordered vowel acoustics is that overall intelligibility, not vowel identification accuracy, has been the perceptual measure of interest. A series of three experiments were conducted to address the problems outlined herein. The goals of the first experiment were to identify subsets of vowel metrics that reliably distinguish speakers with dysarthria from non-disordered speakers and differentiate the dysarthria subtypes. Vowel metrics that capture vowel centralization and reduced spectral distinctiveness among vowels differentiated dysarthric from non-disordered speakers. Vowel metrics generally failed to differentiate speakers according to their dysarthria diagnosis. The second and third experiments were conducted to evaluate the relationship between degraded vowel acoustics and the resulting percept. In the second experiment, correlation and regression analyses revealed vowel metrics that capture vowel centralization and distinctiveness and movement of the second formant frequency were most predictive of vowel identification accuracy and overall intelligibility. The third

  14. Effect of body position on vocal tract acoustics: Acoustic pharyngometry and vowel formants.

    PubMed

    Vorperian, Houri K; Kurtzweil, Sara L; Fourakis, Marios; Kent, Ray D; Tillman, Katelyn K; Austin, Diane

    2015-08-01

    The anatomic basis and articulatory features of speech production are often studied with imaging studies that are typically acquired in the supine body position. It is important to determine if changes in body orientation to the gravitational field alter vocal tract dimensions and speech acoustics. The purpose of this study was to assess the effect of body position (upright versus supine) on (1) oral and pharyngeal measurements derived from acoustic pharyngometry and (2) acoustic measurements of fundamental frequency (F0) and the first four formant frequencies (F1-F4) for the quadrilateral point vowels. Data were obtained for 27 male and female participants, aged 17 to 35 yrs. Acoustic pharyngometry showed a statistically significant effect of body position on volumetric measurements, with smaller values in the supine than upright position, but no changes in length measurements. Acoustic analyses of vowels showed significantly larger values in the supine than upright position for the variables of F0, F3, and the Euclidean distance from the centroid to each corner vowel in the F1-F2-F3 space. Changes in body position affected measurements of vocal tract volume but not length. Body position also affected the aforementioned acoustic variables, but the main vowel formants were preserved.

  15. Effect of body position on vocal tract acoustics: Acoustic pharyngometry and vowel formants

    PubMed Central

    Vorperian, Houri K.; Kurtzweil, Sara L.; Fourakis, Marios; Kent, Ray D.; Tillman, Katelyn K.; Austin, Diane

    2015-01-01

    The anatomic basis and articulatory features of speech production are often studied with imaging studies that are typically acquired in the supine body position. It is important to determine if changes in body orientation to the gravitational field alter vocal tract dimensions and speech acoustics. The purpose of this study was to assess the effect of body position (upright versus supine) on (1) oral and pharyngeal measurements derived from acoustic pharyngometry and (2) acoustic measurements of fundamental frequency (F0) and the first four formant frequencies (F1–F4) for the quadrilateral point vowels. Data were obtained for 27 male and female participants, aged 17 to 35 yrs. Acoustic pharyngometry showed a statistically significant effect of body position on volumetric measurements, with smaller values in the supine than upright position, but no changes in length measurements. Acoustic analyses of vowels showed significantly larger values in the supine than upright position for the variables of F0, F3, and the Euclidean distance from the centroid to each corner vowel in the F1-F2-F3 space. Changes in body position affected measurements of vocal tract volume but not length. Body position also affected the aforementioned acoustic variables, but the main vowel formants were preserved. PMID:26328699

  16. Acoustic characteristics of vowels and plosives/affricates of Mandarin-speaking hearing-impaired children.

    PubMed

    Tseng, Shu-Chuan; Kuei, Ko; Tsou, Pei-Chen

    2011-09-01

    This article presents the results of an acoustic analysis of vowels and plosives/affricates produced by 45 Mandarin-speaking children with hearing impairment. Vowel production is represented and categorized into three groups by vowel space size calculated with normalized F1 and F2 values of corner vowels. The correlation between speech intelligibility and language abilities assessed by the level of word comprehension and the complexity of sentence structure is statistically significant. Vowel space grouping is correlated with speech intelligibility and spike percentage of plosives/affricates production. The generalized linear model analysis also shows that the level of word comprehension and the degree of hearing loss are the two most significant factors in predicting speech intelligibility. The statistical results suggest that the interplay of acoustic characteristic and speech ability is complex.

  17. Revisiting the Canadian English vowel space

    NASA Astrophysics Data System (ADS)

    Hagiwara, Robert

    2005-04-01

    In order to fill a need for experimental-acoustic baseline measurements of Canadian English vowels, a database is currently being constructed in Winnipeg, Manitoba. The database derives from multiple repetitions of fifteen English vowels (eleven standard monophthongs, syllabic /r/ and three standard diphthongs) in /hVd/ and /hVt/ contexts, as spoken by multiple speakers. Frequencies of the first four formants are taken from three timepoints in every vowel token (25, 50, and 75% of vowel duration). Preliminary results (from five men and five women) confirm some features characteristic of Canadian English, but call others into question. For instance the merger of low back vowels appears to be complete for these speakers, but the result is a lower-mid and probably rounded vowel rather than the low back unround vowel often described. With these data Canadian Raising can be quantified as an average 200 Hz or 1.5 Bark downward shift in the frequency of F1 before voiceless /t/. Analysis of the database will lead to a more accurate picture of the Canadian English vowel system, as well as provide a practical and up-to-date point of reference for further phonetic and sociophonetic comparisons.

  18. Effects of deafness on acoustic characteristics of American English tense/lax vowels in maternal speech to infants

    PubMed Central

    Kondaurova, Maria V.; Bergeson, Tonya R.; Dilley, Laura C.

    2012-01-01

    Recent studies have demonstrated that mothers exaggerate phonetic properties of infant-directed (ID) speech. However, these studies focused on a single acoustic dimension (frequency), whereas speech sounds are composed of multiple acoustic cues. Moreover, little is known about how mothers adjust phonetic properties of speech to children with hearing loss. This study examined mothers’ production of frequency and duration cues to the American English tense/lax vowel contrast in speech to profoundly deaf (N = 14) and normal-hearing (N = 14) infants, and to an adult experimenter. First and second formant frequencies and vowel duration of tense (/i/, /u/) and lax (/I/, /ʊ/) vowels were measured. Results demonstrated that for both infant groups mothers hyperarticulated the acoustic vowel space and increased vowel duration in ID speech relative to adult-directed speech. Mean F2 values were decreased for the /u/ vowel and increased for the /I/ vowel, and vowel duration was longer for the /i/, /u/, and /I/ vowels in ID speech. However, neither acoustic cue differed in speech to hearing-impaired or normal-hearing infants. These results suggest that both formant frequencies and vowel duration that differentiate American English tense/lx vowel contrasts are modified in ID speech regardless of the hearing status of the addressee. PMID:22894224

  19. The effect of reduced vowel working space on speech intelligibility in Mandarin-speaking young adults with cerebral palsy

    NASA Astrophysics Data System (ADS)

    Liu, Huei-Mei; Tsao, Feng-Ming; Kuhl, Patricia K.

    2005-06-01

    The purpose of this study was to examine the effect of reduced vowel working space on dysarthric talkers' speech intelligibility using both acoustic and perceptual approaches. In experiment 1, the acoustic-perceptual relationship between vowel working space area and speech intelligibility was examined in Mandarin-speaking young adults with cerebral palsy. Subjects read aloud 18 bisyllabic words containing the vowels /eye/, /aye/, and /you/ using their normal speaking rate. Each talker's words were identified by three normal listeners. The percentage of correct vowel and word identification were calculated as vowel intelligibility and word intelligibility, respectively. Results revealed that talkers with cerebral palsy exhibited smaller vowel working space areas compared to ten age-matched controls. The vowel working space area was significantly correlated with vowel intelligibility (r=0.632, p<0.005) and with word intelligibility (r=0.684, p<0.005). Experiment 2 examined whether tokens of expanded vowel working spaces were perceived as better vowel exemplars and represented with greater perceptual spaces than tokens of reduced vowel working spaces. The results of the perceptual experiment support this prediction. The distorted vowels of talkers with cerebral palsy compose a smaller acoustic space that results in shrunken intervowel perceptual distances for listeners. .

  20. Vowel Acoustics in Dysarthria: Mapping to Perception

    ERIC Educational Resources Information Center

    Lansford, Kaitlin L.; Liss, Julie M.

    2014-01-01

    Purpose: The aim of the present report was to explore whether vowel metrics, demonstrated to distinguish dysarthric and healthy speech in a companion article (Lansford & Liss, 2014), are able to predict human perceptual performance. Method: Vowel metrics derived from vowels embedded in phrases produced by 45 speakers with dysarthria were…

  1. Perceptual assimilation of French and German vowels by American English monolinguals: Acoustic similarity does not predict perceptual similarity

    NASA Astrophysics Data System (ADS)

    Strange, Winifred; Levy, Erika; Lehnholf, Robert

    2001-05-01

    Previous research in our laboratory has demonstrated that the perceived similarity of vowels across languages is not always predictable from the closeness of their target formant values in F1/F2/F3 space. In this study, perceptual similarity was established using a task in which 11 American English (AE) monolinguals were presented multiple tokens of 9 French vowels and 14 North German vowels (in separate blocks) produced in citation-form /hVb(a)/ (bi)syllables by native speakers. They selected 1 of 11 AE vowel responses to which each non-native vowel token was most similar, and rated its goodness on a 9-point Likert scale. Of special interest was the perceptual assimilation of front rounded French [y, oe] and German [y, Y, o/, oe] vowels. Acoustically, all six French and German vowels are more similar to front unrounded AE vowels. However, all six vowels were perceived to be more similar to back rounded AE vowels (range across vowels = 55% to 100%), although relatively poor exemplars. There were differences across languages in how the same vowel was assimilated (e.g., French /y/ assimilated to front AE vowels 13%, German /y/, 0% French [oe] 3%, German [oe] 45%). There were also large individual differences in listeners assimilation patterns. [Work supported by NIDCD.

  2. Variability in English vowels is comparable in articulation and acoustics

    PubMed Central

    Noiray, Aude; Iskarous, Khalil; Whalen, D. H.

    2014-01-01

    The nature of the links between speech production and perception has been the subject of longstanding debate. The present study investigated the articulatory parameter of tongue height and the acoustic F1-F0 difference for the phonological distinction of vowel height in American English front vowels. Multiple repetitions of /i, ɪ, e, ε, æ/ in [(h)Vd] sequences were recorded in seven adult speakers. Articulatory (ultrasound) and acoustic data were collected simultaneously to provide a direct comparison of variability in vowel production in both domains. Results showed idiosyncratic patterns of articulation for contrasting the three front vowel pairs /i-ɪ/, /e-ε/ and /ε-æ/ across subjects, with the degree of variability in vowel articulation comparable to that observed in the acoustics for all seven participants. However, contrary to what was expected, some speakers showed reversals for tongue height for /ɪ/-/e/ that was also reflected in acoustics with F1 higher for /ɪ/ than for /e/. The data suggest the phonological distinction of height is conveyed via speaker-specific articulatory-acoustic patterns that do not strictly match features descriptions. However, the acoustic signal is faithful to the articulatory configuration that generated it, carrying the crucial information for perceptual contrast. PMID:25101144

  3. A cross-dialect acoustic description of vowels: Brazilian and European Portuguese.

    PubMed

    Escudero, Paola; Boersma, Paul; Rauber, Andréia Schurt; Bion, Ricardo A H

    2009-09-01

    This paper examines four acoustic correlates of vowel identity in Brazilian Portuguese (BP) and European Portuguese (EP): first formant (F1), second formant (F2), duration, and fundamental frequency (F0). Both varieties of Portuguese display some cross-linguistically common phenomena: vowel-intrinsic duration, vowel-intrinsic pitch, gender-dependent size of the vowel space, gender-dependent duration, and a skewed symmetry in F1 between front and back vowels. Also, the average difference between the vocal tract sizes associated with /i/ and /u/, as measured from formant analyses, is comparable to the average difference between male and female vocal tract sizes. A language-specific phenomenon is that in both varieties of Portuguese the vowel-intrinsic duration effect is larger than in many other languages. Differences between BP and EP are found in duration (BP has longer stressed vowels than EP), in F1 (the lower-mid front vowel approaches its higher-mid counterpart more closely in EP than in BP), and in the size of the intrinsic pitch effect (larger for BP than for EP).

  4. Acoustic cues to vowel-schwa sequences for high front vowels

    NASA Astrophysics Data System (ADS)

    Slifka, Janet

    2005-09-01

    In a landmark-based model of lexical access [K. N. Stevens, J. Acoust. Soc. Am. 111, 1872-1891 (2002)], the presence of a vowel is marked by a peak in energy in the first formant region. However, when a vowel is followed by a schwa, the schwa frequently appears as a shoulder on the peak associated with the first vowel [W. Howitt, MIT (2000)] two landmarks are not present. The purpose of this study is to examine duration and F2 movement as possible cues to the presence of a vowel-schwa sequence for [+high, +front] vowels. This subset of vowels presents at least two challenges to the detection of a vowel-schwa sequence: (1) duration is expected to contribute to the difference between /i/ and /I/, and (2) an F2 off-glide toward schwa is expected for /I/ in American English. For 613 tokens from the phonetically labeled TIMIT database, equally distributed between /i squflg/, /i/, and /I/, a measure of F2 curvature is a stronger cue than duration in classifying the tokens. Using F2 curvature, over 93% of the tokens are correctly classified. Using duration, 67% are correctly classified. F2 curvature may reflect a more extreme articulation for the /i/ prior to the schwa. [Work supported by NIH Grant DC02978.

  5. Spanish is better than English for discriminating Portuguese vowels: acoustic similarity versus vowel inventory size

    PubMed Central

    Elvin, Jaydene; Escudero, Paola; Vasiliev, Polina

    2014-01-01

    Second language (L2) learners often struggle to distinguish sound contrasts that are not present in their native language (L1). Models of non-native and L2 sound perception claim that perceptual similarity between L1 and L2 sound contrasts correctly predicts discrimination by naïve listeners and L2 learners. The present study tested the explanatory power of vowel inventory size versus acoustic properties as predictors of discrimination accuracy when naïve Australian English (AusE) and Iberian Spanish (IS) listeners are presented with six Brazilian Portuguese (BP) vowel contrasts. Our results show that IS listeners outperformed AusE listeners, confirming that cross-linguistic acoustic properties, rather than cross-linguistic vowel inventory sizes, successfully predict non-native discrimination difficulty. Furthermore, acoustic distance between BP vowels and closest L1 vowels successfully predicted differential levels of difficulty among the six BP contrasts, with BP /e-i/ and /o-u/ being the most difficult for both listener groups. We discuss the importance of our findings for the adequacy of models of L2 speech perception. PMID:25400599

  6. Unmasking the acoustic effects of vowel-to-vowel coarticulation: A statistical modeling approach

    PubMed Central

    Cole, Jennifer; Linebaugh, Gary; Munson, Cheyenne; McMurray, Bob

    2010-01-01

    Coarticulation is a source of acoustic variability for vowels, but how large is this effect relative to other sources of variance? We investigate acoustic effects of anticipatory V-to-V coarticulation relative to variation due to the following C and individual speaker. We examine F1 and F2 from V1 in 48 V1-C#V2 contexts produced by 10 speakers of American English. ANOVA reveals significant effects of both V2 and C on F1 and F2 measures of V1. The influence of V2 and C on acoustic variability relative to that of speaker and target vowel identity is evaluated using hierarchical linear regression. Speaker and target vowel account for roughly 80% of the total variance in F1 and F2, but when this variance is partialed out C and V2 account for another 18% (F1) and 63% (F2) of the remaining target vowel variability. Multinomial logistic regression (MLR) models are constructed to test the power of target vowel F1 and F2 for predicting C and V2 of the upcoming context. Prediction accuracy is 58% for C-Place, 76% for C-Voicing and 54% for V2, but only when variance due to other sources is factored out. MLR is discussed as a model of the parsing mechanism in speech perception. PMID:21173864

  7. The effect of vowel inventory and acoustic properties in Salento Italian learners of Southern British English vowels.

    PubMed

    Escudero, Paola; Sisinni, Bianca; Grimaldi, Mirko

    2014-03-01

    Salento Italian (SI) listeners' categorization and discrimination of standard Southern British English (SSBE) vowels were examined in order to establish their initial state in the acquisition of the SSBE vowel system. The results of the vowel categorization task revealed that SI listeners showed single-category assimilation for many SSBE vowels and multiple-category assimilation for others. Additionally, SI vowel discrimination accuracy varied across contrasts, in line with the categorization results. This differential level of difficulty is discussed on the basis of current L2 perception models. The SI categorization results were then compared to the previously reported data on Peruvian Spanish (PS) listeners. Both SI and PS have a five-vowel inventory and therefore both listener groups were expected to have similar problems when distinguishing SSBE vowel contrasts, but were predicted to have different mappings of SSBE vowels to native categories due to the differences in the acoustic properties of vowels across the two languages. As predicted by the hypothesis that acoustic differences in production lead to a different nonnative perception, the comparison showed that there was large variability in how SSBE vowels are initially mapped to the specific five-vowel inventory. Predictions for differential L2 development across languages are also provided. PMID:24606292

  8. Acoustic and Perceptual Characteristics of Vowels Produced during Simultaneous Communication

    ERIC Educational Resources Information Center

    Schiavetti, Nicholas; Metz, Dale Evan; Whitehead, Robert L.; Brown, Shannon; Borges, Janie; Rivera, Sara; Schultz, Christine

    2004-01-01

    This study investigated the acoustical and perceptual characteristics of vowels in speech produced during simultaneous communication (SC). Twelve normal hearing, experienced sign language users were recorded under SC and speech alone (SA) conditions speaking a set of sentences containing monosyllabic words designed for measurement of vowel…

  9. Diphthongs in the repopulated vowel space

    NASA Astrophysics Data System (ADS)

    Bogacka, Anna

    2005-04-01

    The study examined 8 British English diphthongs produced by Polish learners of English, testing the diphthongs' quality, duration, nasalization, and occurrence of glottal stops before the diphthongs. There were twelve conditions in which the diphthongs were tested: word-initial, word-final, before a voiced obstruent, before a voiceless obstruent, before a nasal consonant, and before a nasal consonant followed by a fricative, and each of these conditions was tested in a stressed and unstressed position. The diphthongs were tested in real words, embedded in sentences, controlled for the stress position, rhythmic units, and length. The sentences were read by 8 female and 8 male Polish learners of English and control subjects. The aim of the phonetic analysis done with Praat, and employing the methodologies used by Flege (1995) for SLA and Peeters (1991) and Jacewicz, Fujimara, and Fox (2003) for diphthongs, is to examine the shape of the restructured vowel space (Liljencrants and Lindblom 1972; Stevens 1989). The approach taken here is termed Vowel Space Repopulation to emphasize that the vowel space of Polish speakers of English is re-structured by new categories in complex ways which are not adequately captured by traditional notions such as ``transfer,'' ``interference,'' or ``interlanguage.''

  10. The articulatory and acoustical characteristics of the ``apical vowels'' in Beijing Mandarin

    NASA Astrophysics Data System (ADS)

    Lee, Wai-Sum

    2005-09-01

    The study investigates the articulatory and acoustical characteristics of the two so-called ``apical vowels'' in Beijing Mandarin, which have been referred to as ``apical anterior vowel'' and ``apical posterior vowel'' by the linguists in China. The ``apical posterior vowel'' has also been described as a retroflex. The results of an EMA (electromagnetic articulograph) analysis show that both vowels are apical, with the tip of tongue approaching the alveolar region for the ``anterior vowel'' and the postalveolar region for the ``posterior vowel.'' The ``posterior vowel'' is pharyngealized, as the body of tongue in particular the posterodorsal portion is pulled backward toward the pharynx. Acoustical data obtained using the CSL4400 speech analysis software show that the two ``apical vowels'' have similar F1 value. The F2 value is slightly larger for the ``posterior vowel'' than ``anterior vowel.'' Thus, the correlation between a larger F2 and the advanced tongue position is not applicable to these ``apical vowels.'' The main difference between the two ``apical vowels'' is in F3, where the value is much smaller for the ``posterior vowel'' than ``anterior vowel.'' It is assumed that the smaller F3 value for the ``posterior vowel'' is due to pharyngealization.

  11. Nonlinearity between acoustics and articulation in Hungarian transparent vowels

    NASA Astrophysics Data System (ADS)

    Benus, Stefan; Kirke, Karen D.; Gafos, Adamantios I.

    2001-05-01

    We present novel results from the acoustic and articulatory investigation of the production of the transparent vowels (TVs) /i/, /i:/, /e:/ in Hungarian (colon denotes length). The acoustic measurements of the front-back distinction (second formant, the difference of the first and second formants [Ladefoged, 1993]) show that the effect of adjacent back vowels on the front quality of the TVs is only weakly significant. The articulatory measurements of the same data, however, show that adjacent back vowels cause highly significant retraction of the tongue body during the production of the front TVs. The significance of this finding lies in its relevance to the relationship between phonetics and phonology. Our results demonstrate that minor phonetic differences in articulation, impossible to access by traditional theory, correlate with full-fledged phonological alternation of suffix selection in Hungarian. Traditional phonological accounts predict no effect of continuous phonetic details on discrete phonological generalizations. This is supported in our acoustic data but contrasts with our articulatory findings. In the paper we propose a dynamic model where phonological transparency is directly related to nonlinearity between acoustics and articulation [Stevens, 1989; Wood, 1979]. [Work supported by NIH.

  12. An Evaluation of Articulatory Working Space Area in Vowel Production of Adults with Down Syndrome

    ERIC Educational Resources Information Center

    Bunton, Kate; Leddy, Mark

    2011-01-01

    Many adolescents and adults with Down syndrome have reduced speech intelligibility. Reasons for this reduction may relate to differences in anatomy and physiology, both of which are important for creating an intelligible speech signal. The purpose of this study was to document acoustic vowel space and articulatory working space for two adult…

  13. The Acoustic Properties of Vowels: A Tool for Improving Articulation and Comprehension of English

    ERIC Educational Resources Information Center

    McCombs, Candalene J.

    2006-01-01

    Correct pronunciation is often a later step in the process of teaching English as a second language. However, a focus on the correct articulation of vowels can significantly improve listening and comprehension skills as well as articulatory skills. Vowels and consonants differ in their acoustic properties. Unlike consonants, vowel sounds are…

  14. Acoustic comparisons of Japanese and English vowels produced by native speakers of Japanese

    NASA Astrophysics Data System (ADS)

    Nishi, Kanae; Akahane-Yamada, Reiko; Kubo, Rieko; Strange, Winifred

    2003-10-01

    This study explored acoustic similarities/differences between Japanese (J) and American English (AE) vowels produced by native J speakers and compared production patterns to their perceptual assimilation of AE vowels [Strange et al., J. Phonetics 26, 311-344 (1998)]. Eight male native J speakers who had served as listeners in Strange et al. produced 18 Japanese (J) vowels (5 long-short pairs, 2 double vowels, and 3 long-short palatalized pairs) and 11 American English (AE) vowels in /hVbopena/ disyllables embedded in a carrier sentence. Acoustical parameters included formant frequencies at syllable midpoint (F1/F2/F3), formant change from 25% to 75% points in syllable (formant change), and vocalic duration. Results of linear discriminant analyses showed rather poor acoustic differentiation of J vowel categories when F1/F2/F3 served as input variables (60% correct classification), which greatly improved when duration and formant change were added. In contrast, correct classification of J speakers' AE vowels using F1/F2/F3 was very poor (66%) and did not improve much when duration and dynamic information were added. J speakers used duration to differentiate long/short AE vowel contrasts except for mid-to-low back vowels; these vowels were perceptually assimilated to a single Japanese vowel, and are very difficult for Japanese listeners to identify.

  15. Acoustic and Perceptual Description of Vowels in a Speaker with Congenital Aglossia

    ERIC Educational Resources Information Center

    McMicken, Betty; Von Berg, Shelley; Iskarous, Khalil

    2012-01-01

    The goals of this study were to (a) compare the vowel space produced by a person with congenital aglossia (PWCA) with a typical vowel space; (b) investigate listeners' intelligibility for single vowels produced by the PWCA, with and without visual information; and (c) determine whether there is a correlation between scores of speech…

  16. Acoustic characteristics of the vowel systems of six regional varieties of American English

    NASA Astrophysics Data System (ADS)

    Clopper, Cynthia G.; Pisoni, David B.; de Jong, Kenneth

    2005-09-01

    Previous research by speech scientists on the acoustic characteristics of American English vowel systems has typically focused on a single regional variety, despite decades of sociolinguistic research demonstrating the extent of regional phonological variation in the United States. In the present study, acoustic measures of duration and first and second formant frequencies were obtained from five repetitions of 11 different vowels produced by 48 talkers representing both genders and six regional varieties of American English. Results revealed consistent variation due to region of origin, particularly with respect to the production of low vowels and high back vowels. The Northern talkers produced shifted low vowels consistent with the Northern Cities Chain Shift, the Southern talkers produced fronted back vowels consistent with the Southern Vowel Shift, and the New England, Midland, and Western talkers produced the low back vowel merger. These findings indicate that the vowel systems of American English are better characterized in terms of the region of origin of the talkers than in terms of a single set of idealized acoustic-phonetic baselines of ``General'' American English and provide benchmark data for six regional varieties.

  17. Acoustic characteristics of the vowel systems of six regional varieties of American English

    PubMed Central

    Clopper, Cynthia G.; Pisoni, David B.; de Jong, Kenneth

    2012-01-01

    Previous research by speech scientists on the acoustic characteristics of American English vowel systems has typically focused on a single regional variety, despite decades of sociolinguistic research demonstrating the extent of regional phonological variation in the United States. In the present study, acoustic measures of duration and first and second formant frequencies were obtained from five repetitions of 11 different vowels produced by 48 talkers representing both genders and six regional varieties of American English. Results revealed consistent variation due to region of origin, particularly with respect to the production of low vowels and high back vowels. The Northern talkers produced shifted low vowels consistent with the Northern Cities Chain Shift, the Southern talkers produced fronted back vowels consistent with the Southern Vowel Shift, and the New England, Midland, and Western talkers produced the low back vowel merger. These findings indicate that the vowel systems of American English are better characterized in terms of the region of origin of the talkers than in terms of a single set of idealized acoustic-phonetic baselines of “General” American English and provide benchmark data for six regional varieties. PMID:16240825

  18. A Comprehensive Three-Dimensional Cortical Map of Vowel Space

    ERIC Educational Resources Information Center

    Scharinger, Mathias; Idsardi, William J.; Poe, Samantha

    2011-01-01

    Mammalian cortex is known to contain various kinds of spatial encoding schemes for sensory information including retinotopic, somatosensory, and tonotopic maps. Tonotopic maps are especially interesting for human speech sound processing because they encode linguistically salient acoustic properties. In this study, we mapped the entire vowel space…

  19. Acoustic Typology of Vowel Inventories and Dispersion Theory: Insights from a Large Cross-Linguistic Corpus

    ERIC Educational Resources Information Center

    Becker-Kristal, Roy

    2010-01-01

    This dissertation examines the relationship between the structural, phonemic properties of vowel inventories and their acoustic phonetic realization, with particular focus on the adequacy of Dispersion Theory, which maintains that inventories are structured so as to maximize perceptual contrast between their component vowels. In order to assess…

  20. Quantitative and Descriptive Comparison of Four Acoustic Analysis Systems: Vowel Measurements

    ERIC Educational Resources Information Center

    Burris, Carlyn; Vorperian, Houri K.; Fourakis, Marios; Kent, Ray D.; Bolt, Daniel M.

    2014-01-01

    Purpose: This study examines accuracy and comparability of 4 trademarked acoustic analysis software packages (AASPs): Praat, WaveSurfer, TF32, and CSL by using synthesized and natural vowels. Features of AASPs are also described. Method: Synthesized and natural vowels were analyzed using each of the AASP's default settings to secure 9…

  1. English vowel spaces produced by Japanese speakers: the smaller point vowels' and the greater schwas'.

    PubMed

    Tomita, Kaoru; Yamada, Jun; Takatsuka, Shigenobu

    2010-10-01

    This study investigated how Japanese-speaking learners of English pronounce the three point vowels /i/, /u/, and /a/ appearing in the first and second monosyllabic words of English noun phrases, and the schwa /ə/ appearing in English disyllabic words. First and second formant (F1 and F2) values were measured for four Japanese speakers and two American English speakers. The hypothesis that the area encompassed by the point vowels in the F1-F2 vowel space tends to be smaller for the Japanese speakers than for the English speakers was verified. The hypothesis that the area formed by the three schwas in chicke_n, spoonfu_l, and Tarza_n is greater for the Japanese speakers than for the English speakers and its related hypothesis were largely upheld. Implications for further research are briefly discussed.

  2. Vowel space development in a child acquiring English and Spanish from birth

    NASA Astrophysics Data System (ADS)

    Andruski, Jean; Kim, Sahyang; Nathan, Geoffrey; Casielles, Eugenia; Work, Richard

    2005-04-01

    To date, research on bilingual first language acquisition has tended to focus on the development of higher levels of language, with relatively few analyses of the acoustic characteristics of bilingual infants' and childrens' speech. Since monolingual infants begin to show perceptual divisions of vowel space that resemble adult native speakers divisions by about 6 months of age [Kuhl et al., Science 255, 606-608 (1992)], bilingual childrens' vowel production may provide evidence of their awareness of language differences relatively early during language development. This paper will examine the development of vowel categories in a child whose mother is a native speaker of Castilian Spanish, and whose father is a native speaker of American English. Each parent speaks to the child only in her/his native language. For this study, recordings made at the ages of 2;5 and 2;10 were analyzed and F1-F2 measurements were made of vowels from the stressed syllables of content words. The development of vowel space is compared across ages within each language, and across languages at each age. In addition, the child's productions are compared with the mother's and father's vocalic productions, which provide the predominant input in Spanish and English respectively.

  3. Vowel Acoustics in Dysarthria: Speech Disorder Diagnosis and Classification

    ERIC Educational Resources Information Center

    Lansford, Kaitlin L.; Liss, Julie M.

    2014-01-01

    Purpose: The purpose of this study was to determine the extent to which vowel metrics are capable of distinguishing healthy from dysarthric speech and among different forms of dysarthria. Method: A variety of vowel metrics were derived from spectral and temporal measurements of vowel tokens embedded in phrases produced by 45 speakers with…

  4. Does Vowel Inventory Density Affect Vowel-to-Vowel Coarticulation?

    ERIC Educational Resources Information Center

    Mok, Peggy P. K.

    2013-01-01

    This study tests the output constraints hypothesis that languages with a crowded phonemic vowel space would allow less vowel-to-vowel coarticulation than languages with a sparser vowel space to avoid perceptual confusion. Mandarin has fewer vowel phonemes than Cantonese, but their allophonic vowel spaces are similarly crowded. The hypothesis…

  5. Acoustic properties of vowels in clear and conversational speech by female non-native English speakers

    NASA Astrophysics Data System (ADS)

    Li, Chi-Nin; So, Connie K.

    2005-04-01

    Studies have shown that talkers can improve the intelligibility of their speech when instructed to speak as if talking to a hearing-impaired person. The improvement of speech intelligibility is associated with specific acoustic-phonetic changes: increases in vowel duration and fundamental frequency (F0), a wider pitch range, and a shift in formant frequencies for F1 and F2. Most previous studies of clear speech production have been conducted with native speakers; research with second language speakers is much less common. The present study examined the acoustic properties of non-native English vowels produced in a clear speaking style. Five female Cantonese speakers and a comparison group of English speakers were recorded producing four vowels (/i u ae a/) in /bVt/ context in conversational and clear speech. Vowel durations, F0, pitch range, and the first two formants for each of the four vowels were measured. Analyses revealed that for both groups of speakers, vowel durations, F0, pitch range, and F1 spoken clearly were greater than those produced conversationally. However, F2 was higher in conversational speech than in clear speech. The findings suggest that female non-native English speakers exhibit acoustic-phonetic patterns similar to those of native speakers when asked to produce English vowels clearly.

  6. A study of high front vowels with articulatory data and acoustic simulations

    PubMed Central

    Jackson, Michel T.-T.; McGowan, Richard S.

    2012-01-01

    The purpose of this study is to test a methodology for describing the articulation of vowels. High front vowels are a test case because some theories suggest that high front vowels have little cross-linguistic variation. Acoustic studies appear to show counterexamples to these predictions, but purely acoustic studies are difficult to interpret because of the many-to-one relation between articulation and acoustics. In this study, vocal tract dimensions, including constriction degree and position, are measured from cinéradiographic and x-ray data on high front vowels from three different languages (North American English, French, and Mandarin Chinese). Statistical comparisons find several significant articulatory differences between North American English /i/ and Mandarin Chinese and French /i/. In particular, differences in constriction degree were found, but not constriction position. Articulatory synthesis is used to model the acoustic consequences of some of the significant articulatory differences, finding that the articulatory differences may have the acoustic consequences of making the latter languages’ /i/ perceptually sharper by shifting the frequencies of F2 and F3 upwards. In addition, the vowel /y/ has specific articulations that differ from those for /i/, including a wider tongue constriction, and substantially different acoustic sensitivity functions for F2 and F3. PMID:22501077

  7. Vowel Acoustics in Adults with Apraxia of Speech

    ERIC Educational Resources Information Center

    Jacks, Adam; Mathes, Katey A.; Marquardt, Thomas P.

    2010-01-01

    Purpose: To investigate the hypothesis that vowel production is more variable in adults with acquired apraxia of speech (AOS) relative to healthy individuals with unimpaired speech. Vowel formant frequency measures were selected as the specific target of focus. Method: Seven adults with AOS and aphasia produced 15 repetitions of 6 American English…

  8. Quantitative and descriptive comparison of four acoustic analysis systems: vowel measurements

    PubMed Central

    Burris, Carlyn; Vorperian, Houri K.; Fourakis, Marios; Kent, Ray D.; Bolt, Daniel M.

    2013-01-01

    Purpose This study examines accuracy and comparability of four trademarked acoustic analysis software packages (AASP): Praat, Wavesurfer, TF32 and CSL using synthesized and natural vowels. Features of AASP are also described. Methods Synthesized and natural vowels were analyzed using each of AASP’s default settings to secure nine acoustic measures: fundamental frequency (F0), formant frequencies (F1-F4), and formant bandwidths (B1-B4). The discrepancy between the software measured values and the input values (synthesized, previously reported, and manual measurements) was used to assess comparability and accuracy. Basic AASP features are described. Results Results indicate that Praat, Wavesurfer, and TF32 generate accurate and comparable F0 and F1-F4 data for synthesized vowels and adult male natural vowels. Results varied by vowel for adult females and children, with some serious errors. Bandwidth measurements by AASPs were highly inaccurate as compared to manual measurements and published data on formant bandwidths. Conclusions Values of F0 and F1-F4 are generally consistent and fairly accurate for adult vowels and for some child vowels using the default settings in Praat, Wavesurfer, and TF32. Manipulation of default settings yields improved output values in TF32 and CSL. Caution is recommended especially before accepting F1-F4 results for children and B1-B4 results for all speakers. PMID:24687465

  9. Functional Connectivity Associated with Acoustic Stability During Vowel Production: Implications for Vocal-Motor Control

    PubMed Central

    2015-01-01

    Abstract Vowels provide the acoustic foundation of communication through speech and song, but little is known about how the brain orchestrates their production. Positron emission tomography was used to study regional cerebral blood flow (rCBF) during sustained production of the vowel /a/. Acoustic and blood flow data from 13, normal, right-handed, native speakers of American English were analyzed to identify CBF patterns that predicted the stability of the first and second formants of this vowel. Formants are bands of resonance frequencies that provide vowel identity and contribute to voice quality. The results indicated that formant stability was directly associated with blood flow increases and decreases in both left- and right-sided brain regions. Secondary brain regions (those associated with the regions predicting formant stability) were more likely to have an indirect negative relationship with first formant variability, but an indirect positive relationship with second formant variability. These results are not definitive maps of vowel production, but they do suggest that the level of motor control necessary to produce stable vowels is reflected in the complexity of an underlying neural system. These results also extend a systems approach to functional image analysis, previously applied to normal and ataxic speech rate that is solely based on identifying patterns of brain activity associated with specific performance measures. Understanding the complex relationships between multiple brain regions and the acoustic characteristics of vocal stability may provide insight into the pathophysiology of the dysarthrias, vocal disorders, and other speech changes in neurological and psychiatric disorders. PMID:25295385

  10. Prosodic domain-initial effects on the acoustic structure of vowels

    NASA Astrophysics Data System (ADS)

    Fox, Robert Allen; Jacewicz, Ewa; Salmons, Joseph

    2003-10-01

    In the process of language change, vowels tend to shift in ``chains,'' leading to reorganizations of entire vowel systems over time. A long research tradition has described such patterns, but little is understood about what factors motivate such shifts. Drawing data from changes in progress in American English dialects, the broad hypothesis is tested that changes in vowel systems are related to prosodic organization and stress patterns. Changes in vowels under greater prosodic prominence correlate directly with, and likely underlie, historical patterns of shift. This study examines acoustic characteristics of vowels at initial edges of prosodic domains [Fougeron and Keating, J. Acoust. Soc. Am. 101, 3728-3740 (1997)]. The investigation is restricted to three distinct prosodic levels: utterance (sentence-initial), phonological phrase (strong branch of a foot), and syllable (weak branch of a foot). The predicted changes in vowels /e/ and /ɛ/ in two American English dialects (from Ohio and Wisconsin) are examined along a set of acoustic parameters: duration, formant frequencies (including dynamic changes over time), and fundamental frequency (F0). In addition to traditional methodology which elicits list-like intonation, a design is adapted to examine prosodic patterns in more typical sentence intonations. [Work partially supported by NIDCD R03 DC005560-01.

  11. Control of spoken vowel acoustics and the influence of phonetic context in human speech sensorimotor cortex.

    PubMed

    Bouchard, Kristofer E; Chang, Edward F

    2014-09-17

    Speech production requires the precise control of vocal tract movements to generate individual speech sounds (phonemes) which, in turn, are rapidly organized into complex sequences. Multiple productions of the same phoneme can exhibit substantial variability, some of which is inherent to control of the vocal tract and its biomechanics, and some of which reflects the contextual effects of surrounding phonemes ("coarticulation"). The role of the CNS in these aspects of speech motor control is not well understood. To address these issues, we recorded multielectrode cortical activity directly from human ventral sensory-motor cortex (vSMC) during the production of consonant-vowel syllables. We analyzed the relationship between the acoustic parameters of vowels (pitch and formants) and cortical activity on a single-trial level. We found that vSMC activity robustly predicted acoustic parameters across vowel categories (up to 80% of variance), as well as different renditions of the same vowel (up to 25% of variance). Furthermore, we observed significant contextual effects on vSMC representations of produced phonemes that suggest active control of coarticulation: vSMC representations for vowels were biased toward the representations of the preceding consonant, and conversely, representations for consonants were biased toward upcoming vowels. These results reveal that vSMC activity for phonemes are not invariant and provide insight into the cortical mechanisms of coarticulation.

  12. Control of Spoken Vowel Acoustics and the Influence of Phonetic Context in Human Speech Sensorimotor Cortex

    PubMed Central

    Bouchard, Kristofer E.

    2014-01-01

    Speech production requires the precise control of vocal tract movements to generate individual speech sounds (phonemes) which, in turn, are rapidly organized into complex sequences. Multiple productions of the same phoneme can exhibit substantial variability, some of which is inherent to control of the vocal tract and its biomechanics, and some of which reflects the contextual effects of surrounding phonemes (“coarticulation”). The role of the CNS in these aspects of speech motor control is not well understood. To address these issues, we recorded multielectrode cortical activity directly from human ventral sensory-motor cortex (vSMC) during the production of consonant-vowel syllables. We analyzed the relationship between the acoustic parameters of vowels (pitch and formants) and cortical activity on a single-trial level. We found that vSMC activity robustly predicted acoustic parameters across vowel categories (up to 80% of variance), as well as different renditions of the same vowel (up to 25% of variance). Furthermore, we observed significant contextual effects on vSMC representations of produced phonemes that suggest active control of coarticulation: vSMC representations for vowels were biased toward the representations of the preceding consonant, and conversely, representations for consonants were biased toward upcoming vowels. These results reveal that vSMC activity for phonemes are not invariant and provide insight into the cortical mechanisms of coarticulation. PMID:25232105

  13. Exploration of Acoustic Features for Automatic Vowel Discrimination in Spontaneous Speech

    ERIC Educational Resources Information Center

    Tyson, Na'im R.

    2012-01-01

    In an attempt to understand what acoustic/auditory feature sets motivated transcribers towards certain labeling decisions, I built machine learning models that were capable of discriminating between canonical and non-canonical vowels excised from the Buckeye Corpus. Specifically, I wanted to model when the dictionary form and the transcribed-form…

  14. Comment on "A geometric representation of spectral and temporal vowel features: quantification of vowel overlap in three linguistic varieties" [J. Acoust. Soc. Am. 119, 2334-2350 (2006)].

    PubMed

    Morrison, Geoffrey Stewart

    2008-01-01

    In a recent paper by Wassink [J. Acoust Soc. Am. 119, 2334-2350 (2006)] the spectral overlap assessment metric (SOAM) was proposed for quantifying the degree of acoustic overlap between vowels. The SOAM does not fully take account of probability densities. An alternative metric is proposed which is based on quadratic discriminant analysis and takes account of probability densities in the form of a posteriori probabilities. Unlike the SOAM, the a posteriori probability-based metric allows for a direct comparison of vowel overlaps calculated using different numbers of dimensions, e.g., three dimensions (Fl, F2, and duration) versus two dimensions (Fl and F2).

  15. Subthalamic Stimulation Reduces Vowel Space at the Initiation of Sustained Production: Implications for Articulatory Motor Control in Parkinson’s Disease

    PubMed Central

    Sidtis, John J.; Alken, Amy G.; Tagliati, Michele; Alterman, Ron; Van Lancker Sidtis, Diana

    2016-01-01

    Background: Stimulation of the subthalamic nuclei (STN) is an effective treatment for Parkinson’s disease, but complaints of speech difficulties after surgery have been difficult to quantify. Speech measures do not convincingly account for such reports. Objective: This study examined STN stimulation effects on vowel production, in order to probe whether DBS affects articulatory posturing. The objective was to compare positioning during the initiation phase with the steady prolongation phase by measuring vowel spaces for three “corner” vowels at these two time frames. Methods: Vowel space was measured over the initial 0.25 sec of sustained productions of high front (/i/), high back (/u/) and low vowels (/a/), and again during a 2 sec segment at the midpoint. Eight right-handed male subjects with bilateral STN stimulation and seven age-matched male controls were studied based on their participation in a larger study that included functional imaging. Mean values: age = 57±4.6 yrs; PD duration = 12.3±2.7 yrs; duration of DBS = 25.6±21.2 mos, and UPDRS III speech score = 1.6±0.7. STN subjects were studied off medication at their therapeutic DBS settings and again with their stimulators off, counter-balanced order. Results: Vowel space was larger in the initiation phase compared to the midpoint for both the control and the STN subjects off stimulation. With stimulation on, however, the initial vowel space was significantly reduced to the area measured at the mid-point. For the three vowels, the acoustics were differentially affected, in accordance with expected effects of front versus back position in the vocal tract. Conclusions: STN stimulation appears to constrain initial articulatory gestures for vowel production, raising the possibility that articulatory positions normally used in speech are similarly constrained. PMID:27003219

  16. Identification of Acoustically Similar and Dissimilar Vowels in Profoundly Deaf Adults Who Use Hearing Aids and/or Cochlear Implants: Some Preliminary Findings

    PubMed Central

    Hay-McCutcheon, Marcia J.; Peterson, Nathaniel R.; Rosado, Christian A.; Pisoni, David B.

    2014-01-01

    Purpose In this study, the authors examined the effects of aging and residual hearing on the identification of acoustically similar and dissimilar vowels in adults with postlingual deafness who use hearing aids (HAs) and/or cochlear implants (CIs). Method The authors used two groups of acoustically similar and dissimilar vowels to assess vowel identification. Also, the Consonant–Nucleus–Consonant Word Recognition Test (Peterson & Lehiste, 1962) and sentences from the Hearing in Noise Test (Nilsson, Soli, & Sullivan, 1994) were administered. Forty CI recipients with postlingual deafness (ages 31–81 years) participated in the study. Results Acoustically similar vowels were more difficult to identify than acoustically dissimilar vowels. With increasing age, performance deteriorated when identifying acoustically similar vowels. Vowel identification was also affected by the use of a contralateral HA and the degree of residual hearing prior to implantation. Moderate correlations were found between speech perception and vowel identification performance. Conclusions Identification performance was affected by the acoustic similarity of the vowels. Older adults experienced more difficulty identifying acoustically similar confusable vowels than did younger adults. The findings might lend support to the ease of language understanding model (Ronnberg, Rudner, Foo, & Lunner, 2008), which proposes that the quality and perceptual robustness of acoustic input affects speech perception. PMID:23824440

  17. A magnetic resonance imaging study on the articulatory and acoustic speech parameters of Malay vowels

    PubMed Central

    2014-01-01

    The phonetic properties of six Malay vowels are investigated using magnetic resonance imaging (MRI) to visualize the vocal tract in order to obtain dynamic articulatory parameters during speech production. To resolve image blurring due to the tongue movement during the scanning process, a method based on active contour extraction is used to track tongue contours. The proposed method efficiently tracks tongue contours despite the partial blurring of MRI images. Consequently, the articulatory parameters that are effectively measured as tongue movement is observed, and the specific shape of the tongue and its position for all six uttered Malay vowels are determined. Speech rehabilitation procedure demands some kind of visual perceivable prototype of speech articulation. To investigate the validity of the measured articulatory parameters based on acoustic theory of speech production, an acoustic analysis based on the uttered vowels by subjects has been performed. As the acoustic speech and articulatory parameters of uttered speech were examined, a correlation between formant frequencies and articulatory parameters was observed. The experiments reported a positive correlation between the constriction location of the tongue body and the first formant frequency, as well as a negative correlation between the constriction location of the tongue tip and the second formant frequency. The results demonstrate that the proposed method is an effective tool for the dynamic study of speech production. PMID:25060583

  18. Effects of semantic predictability and regional dialect on vowel space reduction.

    PubMed

    Clopper, Cynthia G; Pierrehumbert, Janet B

    2008-09-01

    This study explored the interaction between semantic predictability and regional dialect variation in an analysis of speech produced by college-aged female talkers from the Northern, Midland, and Southern dialects of American English. Previous research on the effects of semantic predictability has shown that vowels in high semantic predictability contexts are temporally and spectrally reduced compared to vowels in low semantic predictability contexts. In the current study, an analysis of vowel duration confirmed temporal reduction in the high predictability condition. An analysis of vowel formant structure and vowel space dispersion revealed overall spectral reduction for the Southern talkers. For the Northern talkers, more extreme Northern Cities shifting occurred in the high predictability condition than in the low predictability condition. No effects of semantic predictability were observed for the Midland talkers. These findings suggest an interaction between semantic and indexical factors in vowel reduction processes.

  19. Vowel Space Characteristics of Speech Directed to Children With and Without Hearing Loss

    PubMed Central

    Wieland, Elizabeth A.; Burnham, Evamarie B.; Kondaurova, Maria; Bergeson, Tonya R.

    2015-01-01

    Purpose This study examined vowel characteristics in adult-directed (AD) and infant-directed (ID) speech to children with hearing impairment who received cochlear implants or hearing aids compared with speech to children with normal hearing. Method Mothers' AD and ID speech to children with cochlear implants (Study 1, n = 20) or hearing aids (Study 2, n = 11) was compared with mothers' speech to controls matched on age and hearing experience. The first and second formants of vowels /i/, /ɑ/, and /u/ were measured, and vowel space area and dispersion were calculated. Results In both studies, vowel space was modified in ID compared with AD speech to children with and without hearing loss. Study 1 showed larger vowel space area and dispersion in ID compared with AD speech regardless of infant hearing status. The pattern of effects of ID and AD speech on vowel space characteristics in Study 2 was similar to that in Study 1, but depended partly on children's hearing status. Conclusion Given previously demonstrated associations between expanded vowel space in ID compared with AD speech and enhanced speech perception skills, this research supports a focus on vowel pronunciation in developing intervention strategies for improving speech-language skills in children with hearing impairment. PMID:25658071

  20. Emotions in freely varying and mono-pitched vowels, acoustic and EGG analyses.

    PubMed

    Waaramaa, Teija; Palo, Pertti; Kankare, Elina

    2015-12-01

    Vocal emotions are expressed either by speech or singing. The difference is that in singing the pitch is predetermined while in speech it may vary freely. It was of interest to study whether there were voice quality differences between freely varying and mono-pitched vowels expressed by professional actors. Given their profession, actors have to be able to express emotions both by speech and singing. Electroglottogram and acoustic analyses of emotional utterances embedded in expressions of freely varying vowels [a:], [i:], [u:] (96 samples) and mono-pitched protracted vowels (96 samples) were studied. Contact quotient (CQEGG) was calculated using 35%, 55%, and 80% threshold levels. Three different threshold levels were used in order to evaluate their effects on emotions. Genders were studied separately. The results suggested significant gender differences for CQEGG 80% threshold level. SPL, CQEGG, and F4 were used to convey emotions, but to a lesser degree, when F0 was predetermined. Moreover, females showed fewer significant variations than males. Both genders used more hypofunctional phonation type in mono-pitched utterances than in the expressions with freely varying pitch. The present material warrants further study of the interplay between CQEGG threshold levels and formant frequencies, and listening tests to investigate the perceptual value of the mono-pitched vowels in the communication of emotions.

  1. Emotions in freely varying and mono-pitched vowels, acoustic and EGG analyses.

    PubMed

    Waaramaa, Teija; Palo, Pertti; Kankare, Elina

    2015-12-01

    Vocal emotions are expressed either by speech or singing. The difference is that in singing the pitch is predetermined while in speech it may vary freely. It was of interest to study whether there were voice quality differences between freely varying and mono-pitched vowels expressed by professional actors. Given their profession, actors have to be able to express emotions both by speech and singing. Electroglottogram and acoustic analyses of emotional utterances embedded in expressions of freely varying vowels [a:], [i:], [u:] (96 samples) and mono-pitched protracted vowels (96 samples) were studied. Contact quotient (CQEGG) was calculated using 35%, 55%, and 80% threshold levels. Three different threshold levels were used in order to evaluate their effects on emotions. Genders were studied separately. The results suggested significant gender differences for CQEGG 80% threshold level. SPL, CQEGG, and F4 were used to convey emotions, but to a lesser degree, when F0 was predetermined. Moreover, females showed fewer significant variations than males. Both genders used more hypofunctional phonation type in mono-pitched utterances than in the expressions with freely varying pitch. The present material warrants further study of the interplay between CQEGG threshold levels and formant frequencies, and listening tests to investigate the perceptual value of the mono-pitched vowels in the communication of emotions. PMID:24998780

  2. A note on the acoustic-phonetic characteristics of non-native English vowels produced in noise

    NASA Astrophysics Data System (ADS)

    Li, Chi-Nin; Munro, Murray J.

    2003-10-01

    The Lombard reflex occurs when people unconsciously raise their vocal levels in the presence of loud background noise. Previous work has established that utterances produced in noisy environments exhibit increases in vowel duration and fundamental frequency (F0), and a shift in formant center frequencies for F1 and F2. Most studies of the Lombard reflex have been conducted with native speakers; research with second-language speakers is much less common. The present study examined the effects of the Lombard reflex on foreign-accented English vowel productions. Seven female Cantonese speakers and a comparison group of English speakers were recorded producing three vowels (/i u a/) in /bVt/ context in quiet and in 70 dB of masking noise. Vowel durations, F0, and the first two formants for each of the three vowels were measured. Analyses revealed that vowel durations and F0 were greater in the vowels produced in noise than those produced in quiet in most cases. First formants, but not F2, were consistently higher in Lombard speech than in normal speech. The findings suggest that non-native English speakers exhibit acoustic-phonetic patterns similar to those of native speakers when producing English vowels in noisy conditions.

  3. English Vowel Spaces Produced by Japanese Speakers: The Smaller Point Vowels' and the Greater Schwas'

    ERIC Educational Resources Information Center

    Tomita, Kaoru; Yamada, Jun; Takatsuka, Shigenobu

    2010-01-01

    This study investigated how Japanese-speaking learners of English pronounce the three point vowels /i/, /u/, and /a/ appearing in the first and second monosyllabic words of English noun phrases, and the schwa /[image omitted]/ appearing in English disyllabic words. First and second formant (F1 and F2) values were measured for four Japanese…

  4. Characteristics of the Lax Vowel Space in Dysarthria

    ERIC Educational Resources Information Center

    Tjaden, Kris; Rivera, Deanna; Wilding, Gregory; Turner, Greg S.

    2005-01-01

    It has been hypothesized that lax vowels may be relatively unaffected by dysarthria, owing to the reduced vocal tract shapes required for these phonetic events (G. S. Turner, K. Tjaden, & G. Weismer, 1995). It also has been suggested that lax vowels may be especially susceptible to speech mode effects (M. A. Picheny, N. I. Durlach, & L. D. Braida,…

  5. Phonetic Modification of Vowel Space in Storybook Speech to Infants up to 2 Years of Age

    PubMed Central

    Burnham, Evamarie B.; Wieland, Elizabeth A.; Kondaurova, Maria V.; McAuley, J. Devin; Bergeson, Tonya R.

    2015-01-01

    Purpose A large body of literature has indicated vowel space area expansion in infant-directed (ID) speech compared with adult-directed (AD) speech, which may promote language acquisition. The current study tested whether this expansion occurs in storybook speech read to infants at various points during their first 2 years of life. Method In 2 studies, mothers read a storybook containing target vowels in ID and AD speech conditions. Study 1 was longitudinal, with 11 mothers recorded when their infants were 3, 6, and 9 months old. Study 2 was cross-sectional, with 48 mothers recorded when their infants were 3, 9, 13, or 20 months old (n = 12 per group). The 1st and 2nd formants of vowels /i/, /ɑ/, and /u/ were measured, and vowel space area and dispersion were calculated. Results Across both studies, 1st and/or 2nd formant frequencies shifted systematically for /i/ and /u/ vowels in ID compared with AD speech. No difference in vowel space area or dispersion was found. Conclusions The results suggest that a variety of communication and situational factors may affect phonetic modifications in ID speech, but that vowel space characteristics in speech to infants stay consistent across the first 2 years of life. PMID:25659121

  6. Vowel Acoustics in Parkinson's Disease and Multiple Sclerosis: Comparison of Clear, Loud, and Slow Speaking Conditions

    ERIC Educational Resources Information Center

    Tjaden, Kris; Lam, Jennifer; Wilding, Greg

    2013-01-01

    Purpose: The impact of clear speech, increased vocal intensity, and rate reduction on acoustic characteristics of vowels was compared in speakers with Parkinson's disease (PD), speakers with multiple sclerosis (MS), and healthy controls. Method: Speakers read sentences in habitual, clear, loud, and slow conditions. Variations in clarity,…

  7. How Native Do They Sound? An Acoustic Analysis of the Spanish Vowels of Elementary Spanish Immersion Students

    ERIC Educational Resources Information Center

    Menke, Mandy R.

    2015-01-01

    Language immersion students' lexical, syntactic, and pragmatic competencies are well documented, yet their phonological skill has remained relatively unexplored. This study investigates the Spanish vowel productions of a cross-sectional sample of 35 one-way Spanish immersion students. Learner productions were analyzed acoustically and compared to…

  8. Acoustic correlates of vowel intelligibility in clear and conversational speech for young normal-hearing and elderly hearing-impaired listeners.

    PubMed

    Ferguson, Sarah Hargus; Quené, Hugo

    2014-06-01

    The present investigation carried out acoustic analyses of vowels in clear and conversational speech produced by 41 talkers. Mixed-effects models were then deployed to examine relationships among acoustic and perceptual data for these vowels. Acoustic data include vowel duration, steady-state formant frequencies, and two measures of dynamic formant movement. Perceptual data consist of vowel intelligibility in noise for young normal-hearing and elderly hearing-impaired listeners, as reported by Ferguson in 2004 and 2012 [J. Acoust. Soc. Am. 116, 2365-2373 (2004); J. Speech Lang. Hear. Res. 55, 779-790 (2012)], respectively. Significant clear speech effects were observed for all acoustic metrics, although not all measures changed for all vowels and considerable talker variability was observed. Mixed-effects analyses revealed that the contribution of duration and steady-state formant information to vowel intelligibility differed for the two listener groups. This outcome is consistent with earlier research suggesting that hearing loss, and possibly aging, alters the way acoustic cues are used for identifying vowels.

  9. Recognition of emotions in Mexican Spanish speech: an approach based on acoustic modelling of emotion-specific vowels.

    PubMed

    Caballero-Morales, Santiago-Omar

    2013-01-01

    An approach for the recognition of emotions in speech is presented. The target language is Mexican Spanish, and for this purpose a speech database was created. The approach consists in the phoneme acoustic modelling of emotion-specific vowels. For this, a standard phoneme-based Automatic Speech Recognition (ASR) system was built with Hidden Markov Models (HMMs), where different phoneme HMMs were built for the consonants and emotion-specific vowels associated with four emotional states (anger, happiness, neutral, sadness). Then, estimation of the emotional state from a spoken sentence is performed by counting the number of emotion-specific vowels found in the ASR's output for the sentence. With this approach, accuracy of 87-100% was achieved for the recognition of emotional state of Mexican Spanish speech.

  10. Recognition of Emotions in Mexican Spanish Speech: An Approach Based on Acoustic Modelling of Emotion-Specific Vowels

    PubMed Central

    Caballero-Morales, Santiago-Omar

    2013-01-01

    An approach for the recognition of emotions in speech is presented. The target language is Mexican Spanish, and for this purpose a speech database was created. The approach consists in the phoneme acoustic modelling of emotion-specific vowels. For this, a standard phoneme-based Automatic Speech Recognition (ASR) system was built with Hidden Markov Models (HMMs), where different phoneme HMMs were built for the consonants and emotion-specific vowels associated with four emotional states (anger, happiness, neutral, sadness). Then, estimation of the emotional state from a spoken sentence is performed by counting the number of emotion-specific vowels found in the ASR's output for the sentence. With this approach, accuracy of 87–100% was achieved for the recognition of emotional state of Mexican Spanish speech. PMID:23935410

  11. Phonetic Modification of Vowel Space in Storybook Speech to Infants up to 2 Years of Age

    ERIC Educational Resources Information Center

    Burnham, Evamarie B.; Wieland, Elizabeth A.; Kondaurova, Maria V.; McAuley, J. Devin; Bergeson, Tonya R.; Dilley, Laura C.

    2015-01-01

    Purpose: A large body of literature has indicated vowel space area expansion in infant-directed (ID) speech compared with adult-directed (AD) speech, which may promote language acquisition. The current study tested whether this expansion occurs in storybook speech read to infants at various points during their first 2 years of life. Method: In 2…

  12. Pitch (F0) and formant profiles of human vowels and vowel-like baboon grunts: The role of vocalizer body size and voice-acoustic allometry

    NASA Astrophysics Data System (ADS)

    Rendall, Drew; Kollias, Sophie; Ney, Christina; Lloyd, Peter

    2005-02-01

    Key voice features-fundamental frequency (F0) and formant frequencies-can vary extensively between individuals. Much of the variation can be traced to differences in the size of the larynx and vocal-tract cavities, but whether these differences in turn simply reflect differences in speaker body size (i.e., neutral vocal allometry) remains unclear. Quantitative analyses were therefore undertaken to test the relationship between speaker body size and voice F0 and formant frequencies for human vowels. To test the taxonomic generality of the relationships, the same analyses were conducted on the vowel-like grunts of baboons, whose phylogenetic proximity to humans and similar vocal production biology and voice acoustic patterns recommend them for such comparative research. For adults of both species, males were larger than females and had lower mean voice F0 and formant frequencies. However, beyond this, F0 variation did not track body-size variation between the sexes in either species, nor within sexes in humans. In humans, formant variation correlated significantly with speaker height but only in males and not in females. Implications for general vocal allometry are discussed as are implications for speech origins theories, and challenges to them, related to laryngeal position and vocal tract length. .

  13. A Neural Substrate for Rapid Timbre Recognition? Neural and Behavioral Discrimination of Very Brief Acoustic Vowels.

    PubMed

    Occelli, F; Suied, C; Pressnitzer, D; Edeline, J-M; Gourévitch, B

    2016-06-01

    The timbre of a sound plays an important role in our ability to discriminate between behaviorally relevant auditory categories, such as different vowels in speech. Here, we investigated, in the primary auditory cortex (A1) of anesthetized guinea pigs, the neural representation of vowels with impoverished timbre cues. Five different vowels were presented with durations ranging from 2 to 128 ms. A psychophysical experiment involving human listeners showed that identification performance was near ceiling for the longer durations and degraded close to chance level for the shortest durations. This was likely due to spectral splatter, which reduced the contrast between the spectral profiles of the vowels at short durations. Effects of vowel duration on cortical responses were well predicted by the linear frequency responses of A1 neurons. Using mutual information, we found that auditory cortical neurons in the guinea pig could be used to reliably identify several vowels for all durations. Information carried by each cortical site was low on average, but the population code was accurate even for durations where human behavioral performance was poor. These results suggest that a place population code is available at the level of A1 to encode spectral profile cues for even very short sounds. PMID:25947234

  14. Effects of Long-Term Tracheostomy on Spectral Characteristics of Vowel Production.

    ERIC Educational Resources Information Center

    Kamen, Ruth Saletsky; Watson, Ben C.

    1991-01-01

    Eight preschool children who underwent tracheotomy during the prelingual period were compared to matched controls on a variety of speech measures. Children with tracheotomies showed reduced acoustic vowel space, suggesting they were limited in their ability to produce extreme vocal tract configurations for vowels postdecannulation. Oral motor…

  15. The influence of sexual orientation on vowel production (L)

    NASA Astrophysics Data System (ADS)

    Pierrehumbert, Janet B.; Bent, Tessa; Munson, Benjamin; Bradlow, Ann R.; Bailey, J. Michael

    2004-10-01

    Vowel production in gay, lesbian, bisexual (GLB), and heterosexual speakers was examined. Differences in the acoustic characteristics of vowels were found as a function of sexual orientation. Lesbian and bisexual women produced less fronted /u/ and /opena/ than heterosexual women. Gay men produced a more expanded vowel space than heterosexual men. However, the vowels of GLB speakers were not generally shifted toward vowel patterns typical of the opposite sex. These results are inconsistent with the conjecture that innate biological factors have a broadly feminizing influence on the speech of gay men and a broadly masculinizing influence on the speech of lesbian/bisexual women. They are consistent with the idea that innate biological factors influence GLB speech patterns indirectly by causing selective adoption of certain speech patterns characteristic of the opposite sex. .

  16. Effects of Age on Concurrent Vowel Perception in Acoustic and Simulated Electroacoustic Hearing

    ERIC Educational Resources Information Center

    Arehart, Kathryn H.; Souza, Pamela E.; Muralimanohar, Ramesh Kumar; Miller, Christi Wise

    2011-01-01

    Purpose: In this study, the authors investigated the effects of age on the use of fundamental frequency differences([delta]F[subscript 0]) in the perception of competing synthesized vowels in simulations of electroacoustic and cochlear-implant hearing. Method: Twelve younger listeners with normal hearing and 13 older listeners with (near) normal…

  17. The Effects of Inventory on Vowel Perception in French and Spanish: An MEG Study

    ERIC Educational Resources Information Center

    Hacquard, Valentine; Walter, Mary Ann; Marantz, Alec

    2007-01-01

    Production studies have shown that speakers of languages with larger phoneme inventories expand their acoustic space relative to languages with smaller inventories [Bradlow, A. (1995). A comparative acoustic study of English and Spanish vowels. "Journal of the Acoustical Society of America," 97(3), 1916-1924; Jongman, A., Fourakis, M., & Sereno,…

  18. Durations of American English vowels by native and non-native speakers: acoustic analyses and perceptual effects.

    PubMed

    Liu, Chang; Jin, Su-Hyun; Chen, Chia-Tsen

    2014-06-01

    The goal of this study was to examine durations of American English vowels produced by English-, Chinese-, and Korean-native speakers and the effects of vowel duration on vowel intelligibility. Twelve American English vowels were recorded in the /hVd/ phonetic context by native speakers and non-native speakers. The English vowel duration patterns as a function of vowel produced by non-native speakers were generally similar to those produced by native speakers. These results imply that using duration differences across vowels may be an important strategy for non-native speakers' production before they are able to employ spectral cues to produce and perceive English speech sounds. In the intelligibility experiment, vowels were selected from 10 native and non-native speakers and vowel durations were equalized at 170 ms. Intelligibility of vowels with original and equalized durations was evaluated by American English native listeners. Results suggested that vowel intelligibility of native and non-native speakers degraded slightly by 3-8% when durations were equalized, indicating that vowel duration plays a minor role in vowel intelligibility.

  19. Acoustic correlates of caller identity and affect intensity in the vowel-like grunt vocalizations of baboons.

    PubMed

    Rendall, Drew

    2003-06-01

    Comparative, production-based research on animal vocalizations can allow assessments of continuity in vocal communication processes across species, including humans, and may aid in the development of general frameworks relating specific constitutional attributes of callers to acoustic-structural details of their vocal output. Analyses were undertaken on vowel-like baboon grunts to examine variation attributable to caller identity and the intensity of the affective state underlying call production. Six hundred six grunts from eight adult females were analyzed. Grunts derived from 128 bouts of calling in two behavioral contexts: concerted group movements and social interactions involving mothers and their young infants. Each context was subdivided into a high- and low-arousal condition. Thirteen acoustic features variously predicted to reflect variation in either caller identity or arousal intensity were measured for each grunt bout, including tempo-, source- and filter-related features. Grunt bouts were highly individually distinctive, differing in a variety of acoustic dimensions but with some indication that filter-related features contributed disproportionately to individual distinctiveness. In contrast, variation according to arousal condition was associated primarily with tempo- and source-related features, many matching those identified as vehicles of affect expression in other nonhuman primate species and in human speech and other nonverbal vocal signals.

  20. Acoustic correlates of caller identity and affect intensity in the vowel-like grunt vocalizations of baboons

    NASA Astrophysics Data System (ADS)

    Rendall, Drew

    2003-06-01

    Comparative, production-based research on animal vocalizations can allow assessments of continuity in vocal communication processes across species, including humans, and may aid in the development of general frameworks relating specific constitutional attributes of callers to acoustic-structural details of their vocal output. Analyses were undertaken on vowel-like baboon grunts to examine variation attributable to caller identity and the intensity of the affective state underlying call production. Six hundred six grunts from eight adult females were analyzed. Grunts derived from 128 bouts of calling in two behavioral contexts: concerted group movements and social interactions involving mothers and their young infants. Each context was subdivided into a high- and low-arousal condition. Thirteen acoustic features variously predicted to reflect variation in either caller identity or arousal intensity were measured for each grunt bout, including tempo-, source- and filter-related features. Grunt bouts were highly individually distinctive, differing in a variety of acoustic dimensions but with some indication that filter-related features contributed disproportionately to individual distinctiveness. In contrast, variation according to arousal condition was associated primarily with tempo- and source-related features, many matching those identified as vehicles of affect expression in other nonhuman primate species and in human speech and other nonverbal vocal signals.

  1. A longitudinal study of very young children's vowel production

    PubMed Central

    McGowan, Rebecca W.; McGowan, Richard S.; Denny, Margaret; Nittrouer, Susan

    2014-01-01

    Purpose Ecologically realistic, spontaneous adult-directed longitudinal speech data of young children was described by acoustic analyses. Method The first two formant frequencies of vowels produced by six children from different American English dialect regions were analyzed from ages 18 to 48 months. The vowels were from largely conversational contexts and were classified according to dictionary pronunciation. Results Within-subject formant frequency variability remained relatively constant for the span of ages studied here. It was often difficult to detect overall decreases in the first two formant frequencies between the ages of 30 and 48 months. A study of the movement of the corner vowels with respect to the vowel centroid showed that the shape of the vowel space remained qualitatively constant from 30 through 48 months. Conclusions The shape of the vowel space is established early in life. Some aspects of regional dialect were observed in some of the subjects at 42 months of age. The present paper adds to the existing data on the development of vowel spaces by describing ecologically realistic speech. PMID:24687464

  2. Articulatory Changes in Muscle Tension Dysphonia: Evidence of Vowel Space Expansion Following Manual Circumlaryngeal Therapy

    ERIC Educational Resources Information Center

    Roy, Nelson; Nissen, Shawn L.; Dromey, Christopher; Sapir, Shimon

    2009-01-01

    In a preliminary study, we documented significant changes in formant transitions associated with successful manual circumlaryngeal treatment (MCT) of muscle tension dysphonia (MTD), suggesting improvement in speech articulation. The present study explores further the effects of MTD on vowel articulation by means of additional vowel acoustic…

  3. The auditory N1m reveals the left-hemispheric representation of vowel identity in humans.

    PubMed

    Mäkelä, Anna Mari; Alku, Paavo; Tiitinen, Hannu

    2003-12-19

    The cortical correlates of the perception of the sustained vowels /a/, /o/ and /u/ were studied by using whole-head magnetoencephalography (MEG). The three vowels which were located on a line in the space spanned by the first (F1) and second (F2) formants and having equal F2-F1 differences evoked equally strong auditory N1m responses at 120 ms after stimulus onset. The left-hemispheric distribution of the source locations, estimated by equivalent current dipoles, reflected the acoustic similarity of the vowels: the growing distance of the vowels in the F2,F1-space was accompanied by a growing distance between the centres of gravity of activation elicited by each vowel. Thus, direct evidence for the orderly left-hemispheric representation of phonemes in human auditory cortex was found.

  4. Vowel Development in an Emergent Mandarin-English Bilingual Child: A Longitudinal Study

    ERIC Educational Resources Information Center

    Yang, Jing; Fox, Robert A.; Jacewicz, Ewa

    2015-01-01

    This longitudinal case study documents the emergence of bilingualism in a young monolingual Mandarin boy on the basis of an acoustic analysis of his vowel productions recorded via a picture-naming task over 20 months following his enrollment in an all-English (L2) preschool at the age of 3;7. The study examined (1) his initial L2 vowel space, (2)…

  5. Thresholds for vowel formant discrimination using a sentence classification task

    NASA Astrophysics Data System (ADS)

    Kewley-Port, Diane; Oglesbee, Eric; Lee, Jae Hee

    2005-09-01

    Accurate classification of vowels in sentences is challenging because American English has many acoustically similar vowels. Using a 2AFC paradigm, our previous research estimated thresholds for vowel formant discrimination in sentences that were two times smaller than the measured formant distance between close vowels. A new paradigm has been developed to estimate the ability to detect formant differences in a sentence classification task. A seven-token continuum of changes in either F1 or F2 was generated from natural productions of ``bid'' and ``bed'' using the synthesizer STRAIGHT. These tokens were spliced into a nine-word sentence at different positions that also contained two other test words, one each from pairs ``cot/cut'' and ``hack/hawk.'' Listeners were asked to identify the three words they heard in the sentence. Listeners also identified whether ``bid'' or ``bed'' was heard when only the isolated tokens were presented. Thresholds to detect a change from ``bid'' were obtained from psychometric functions fit to the data. Thresholds were similar for the sentence and word-only tasks. Overall, thresholds in both classification tasks were worse than those from the 2AFC tasks. Results will be discussed in terms of the relation between these discrimination thresholds, vowel identification, and vowel spaces. [Work supported by NIHDCD-02229.

  6. Vowel acquisition by prelingually deaf children with cochlear implants

    NASA Astrophysics Data System (ADS)

    Bouchard, Marie-Eve; Le Normand, Marie-Thérèse; Ménard, Lucie; Goud, Marilyne; Cohen, Henri

    2001-05-01

    Phonetic transcriptions (study 1) and acoustic analysis (study 2) were used to clarify the nature and rhythm of vowel acquisition following the cochlear implantation of prelingually deaf children. In the first study, seven children were divided according to their degree of hearing loss (DHL): DHL I: 90-100 dB of hearing loss, 1 children; DHL II: 100-110 dB, 3 children; and DHL III: over 110 dB, 3 children. Spontaneous speech productions were recorded and videotaped 6 and 12 months postsurgery and vowel inventories were obtained by listing all vowels that occurred at least twice in the child's repertoire at the time of recording. Results showed that degree of hearing loss and age at implantation have a significant impact on vowel acquisition. Indeed, DHL I and II children demonstrated more diversified as well as more typical pattern of acquisition. In the second study, the values of the first and second formants were extracted. The results suggest evolving use of the acoustic space, reflecting the use of auditory feedback to produce the three phonological features exploited to contrast French vowels (height, place of articulation, and rounding). The possible influence of visual feedback before cochlear implant is discussed.

  7. The Acoustic Characteristics of Diphthongs in Indian English

    ERIC Educational Resources Information Center

    Maxwell, Olga; Fletcher, Janet

    2010-01-01

    This paper presents the results of an acoustic analysis of English diphthongs produced by three L1 speakers of Hindi and four L1 speakers of Punjabi. Formant trajectories of rising and falling diphthongs (i.e., vowels where there is a clear rising or falling trajectory through the F1/F2 vowel space) were analysed in a corpus of citation-form…

  8. A cross-language study of vowel sounds produced with and without emphasis: Testing the theory of adaptive dispersion

    NASA Astrophysics Data System (ADS)

    Hay, Jessica F.; Sato, Momoko; Coren, Amy E.; Diehl, Randy L.

    2002-05-01

    According to Lindbloms Theory of Adaptive Dispersion (TAD), the aim of talkers is to make phonological contrasts sufficiently distinctive to promote linguistic comprehension by the listener while minimizing the articulatory effort needed to achieve this degree of distinctiveness. When part of an utterance carries new-rather than given-information, it tends to be spoken with greater emphasis and clarity. In this study, several possible acoustic correlates of vowels in emphasized words were examined in American English, French, and Japanese in comparable phonetic and sentence contexts. These possible correlates include an expanded vowel space, greater vowel inherent spectral change, and a greater systematic variation in vowel length. Preliminary analyses suggest that the contrast-enhancing properties of emphasized vowels vary considerably across languages. [Work supported by NIDCD.

  9. Malaysian English: An Instrumental Analysis of Vowel Contrasts

    ERIC Educational Resources Information Center

    Pillai, Stefanie; Don, Zuraidah Mohd.; Knowles, Gerald; Tang, Jennifer

    2010-01-01

    This paper makes an instrumental analysis of English vowel monophthongs produced by 47 female Malaysian speakers. The focus is on the distribution of Malaysian English vowels in the vowel space, and the extent to which there is phonetic contrast between traditionally paired vowels. The results indicate that, like neighbouring varieties of English,…

  10. International Space Station Acoustics - A Status Report

    NASA Technical Reports Server (NTRS)

    Allen, Christopher S.

    2015-01-01

    It is important to control acoustic noise aboard the International Space Station (ISS) to provide a satisfactory environment for voice communications, crew productivity, alarm audibility, and restful sleep, and to minimize the risk for temporary and permanent hearing loss. Acoustic monitoring is an important part of the noise control process on ISS, providing critical data for trend analysis, noise exposure analysis, validation of acoustic analyses and predictions, and to provide strong evidence for ensuring crew health and safety, thus allowing Flight Certification. To this purpose, sound level meter (SLM) measurements and acoustic noise dosimetry are routinely performed. And since the primary noise sources on ISS include the environmental control and life support system (fans and airflow) and active thermal control system (pumps and water flow), acoustic monitoring will reveal changes in hardware noise emissions that may indicate system degradation or performance issues. This paper provides the current acoustic levels in the ISS modules and sleep stations and is an update to the status presented in 2011. Since this last status report, many payloads (science experiment hardware) have been added and a significant number of quiet ventilation fans have replaced noisier fans in the Russian Segment. Also, noise mitigation efforts are planned to reduce the noise levels of the T2 treadmill and levels in Node 3, in general. As a result, the acoustic levels on the ISS continue to improve.

  11. International Space Station Acoustics - A Status Report

    NASA Technical Reports Server (NTRS)

    Allen, Christopher S.; Denham, Samuel A.

    2011-01-01

    It is important to control acoustic noise aboard the International Space Station (ISS) to provide a satisfactory environment for voice communications, crew productivity, and restful sleep, and to minimize the risk for temporary and permanent hearing loss. Acoustic monitoring is an important part of the noise control process on ISS, providing critical data for trend analysis, noise exposure analysis, validation of acoustic analysis and predictions, and to provide strong evidence for ensuring crew health and safety, thus allowing Flight Certification. To this purpose, sound level meter (SLM) measurements and acoustic noise dosimetry are routinely performed. And since the primary noise sources on ISS include the environmental control and life support system (fans and airflow) and active thermal control system (pumps and water flow), acoustic monitoring will indicate changes in hardware noise emissions that may indicate system degradation or performance issues. This paper provides the current acoustic levels in the ISS modules and sleep stations, and is an update to the status presented in 20031. Many new modules, and sleep stations have been added to the ISS since that time. In addition, noise mitigation efforts have reduced noise levels in some areas. As a result, the acoustic levels on the ISS have improved.

  12. Shifting Perceptual Weights in L2 Vowel Identification after Training

    PubMed Central

    Hu, Wei; Mi, Lin; Yang, Zhen; Tao, Sha; Li, Mingshuang; Wang, Wenjing; Dong, Qi; Liu, Chang

    2016-01-01

    Difficulties with second-language vowel perception may be related to the significant challenges in using acoustic-phonetic cues. This study investigated the effects of perception training with duration-equalized vowels on native Chinese listeners’ English vowel perception and their use of acoustic-phonetic cues. Seventeen native Chinese listeners were perceptually trained with duration-equalized English vowels, and another 17 native Chinese listeners watched English videos as a control group. Both groups were tested with English vowel identification and vowel formant discrimination before training, immediately after training, and three months later. The results showed that the training effect was greater for the vowel training group than for the control group, while both groups improved their English vowel identification and vowel formant discrimination after training. Moreover, duration-equalized vowel perception training significantly reduced listeners’ reliance on duration cues and improved their use of spectral cues in identifying English vowels, but video-watching did not help. The results suggest that duration-equalized English vowel perception training may improve non-native listeners’ English vowel perception by changing their perceptual weights of acoustic-phonetic cues. PMID:27649413

  13. Shifting Perceptual Weights in L2 Vowel Identification after Training.

    PubMed

    Hu, Wei; Mi, Lin; Yang, Zhen; Tao, Sha; Li, Mingshuang; Wang, Wenjing; Dong, Qi; Liu, Chang

    2016-01-01

    Difficulties with second-language vowel perception may be related to the significant challenges in using acoustic-phonetic cues. This study investigated the effects of perception training with duration-equalized vowels on native Chinese listeners' English vowel perception and their use of acoustic-phonetic cues. Seventeen native Chinese listeners were perceptually trained with duration-equalized English vowels, and another 17 native Chinese listeners watched English videos as a control group. Both groups were tested with English vowel identification and vowel formant discrimination before training, immediately after training, and three months later. The results showed that the training effect was greater for the vowel training group than for the control group, while both groups improved their English vowel identification and vowel formant discrimination after training. Moreover, duration-equalized vowel perception training significantly reduced listeners' reliance on duration cues and improved their use of spectral cues in identifying English vowels, but video-watching did not help. The results suggest that duration-equalized English vowel perception training may improve non-native listeners' English vowel perception by changing their perceptual weights of acoustic-phonetic cues.

  14. Shifting Perceptual Weights in L2 Vowel Identification after Training.

    PubMed

    Hu, Wei; Mi, Lin; Yang, Zhen; Tao, Sha; Li, Mingshuang; Wang, Wenjing; Dong, Qi; Liu, Chang

    2016-01-01

    Difficulties with second-language vowel perception may be related to the significant challenges in using acoustic-phonetic cues. This study investigated the effects of perception training with duration-equalized vowels on native Chinese listeners' English vowel perception and their use of acoustic-phonetic cues. Seventeen native Chinese listeners were perceptually trained with duration-equalized English vowels, and another 17 native Chinese listeners watched English videos as a control group. Both groups were tested with English vowel identification and vowel formant discrimination before training, immediately after training, and three months later. The results showed that the training effect was greater for the vowel training group than for the control group, while both groups improved their English vowel identification and vowel formant discrimination after training. Moreover, duration-equalized vowel perception training significantly reduced listeners' reliance on duration cues and improved their use of spectral cues in identifying English vowels, but video-watching did not help. The results suggest that duration-equalized English vowel perception training may improve non-native listeners' English vowel perception by changing their perceptual weights of acoustic-phonetic cues. PMID:27649413

  15. Some Consequences of Velarization on Catalan Vowels.

    ERIC Educational Resources Information Center

    Widdison, Kirk

    The acoustic effects of the syllable-final /l/ significantly alter the vocalic timbre of the preceding vowel in Catalan. Vowel quality is modified anticipatory to the articulatory gestures required by the /l/, resulting in a lowered second formant. Syllable-final /l/ in Catalan is heavily velarized as a result of tongue tip-tongue back coupling…

  16. English vowel learning by speakers of Mandarin

    NASA Astrophysics Data System (ADS)

    Thomson, Ron I.

    2005-04-01

    One of the most influential models of second language (L2) speech perception and production [Flege, Speech Perception and Linguistic Experience (York, Baltimore, 1995) pp. 233-277] argues that during initial stages of L2 acquisition, perceptual categories sharing the same or nearly the same acoustic space as first language (L1) categories will be processed as members of that L1 category. Previous research has generally been limited to testing these claims on binary L2 contrasts, rather than larger portions of the perceptual space. This study examines the development of 10 English vowel categories by 20 Mandarin L1 learners of English. Imitation of English vowel stimuli by these learners, at 6 data collection points over the course of one year, were recorded. Using a statistical pattern recognition model, these productions were then assessed against native speaker norms. The degree to which the learners' perception/production shifted toward the target English vowels and the degree to which they matched L1 categories in ways predicted by theoretical models are discussed. The results of this experiment suggest that previous claims about perceptual assimilation of L2 categories to L1 categories may be too strong.

  17. Language dependent vowel representation in speech production

    PubMed Central

    Mitsuya, Takashi; Samson, Fabienne; Ménard, Lucie; Munhall, Kevin G.

    2013-01-01

    The representation of speech goals was explored using an auditory feedback paradigm. When talkers produce vowels the formant structure of which is perturbed in real time, they compensate to preserve the intended goal. When vowel formants are shifted up or down in frequency, participants change the formant frequencies in the opposite direction to the feedback perturbation. In this experiment, the specificity of vowel representation was explored by examining the magnitude of vowel compensation when the second formant frequency of a vowel was perturbed for speakers of two different languages (English and French). Even though the target vowel was the same for both language groups, the pattern of compensation differed. French speakers compensated to smaller perturbations and made larger compensations overall. Moreover, French speakers modified the third formant in their vowels to strengthen the compensation even though the third formant was not perturbed. English speakers did not alter their third formant. Changes in the perceptual goodness ratings by the two groups of participants were consistent with the threshold to initiate vowel compensation in production. These results suggest that vowel goals not only specify the quality of the vowel but also the relationship of the vowel to the vowel space of the spoken language. PMID:23654403

  18. Acoustic levitation and manipulation for space applications

    NASA Technical Reports Server (NTRS)

    Wang, T. G.

    1979-01-01

    A wide spectrum of experiments to be performed in space in a microgravity environment require levitation and manipulation of liquid or molten samples. A novel acoustic method has been developed at JPL for controlling liquid samples without physical contacts. This method utilizes the static pressure generated by three orthogonal acoustic standing waves excited within an enclosure. Furthermore, this method will allow the sample to be rotated and/or oscillated by modifying the phase angles and/or the amplitude of the acoustic field. This technique has been proven both in our laboratory and in a microgravity environment provided by KC-135 flights. Samples placed within our chamber driven at (1,0,0), (0,1,0), and (0,0,1), modes were indeed levitated, rotated, and oscillated.

  19. Learning to pronounce Vowel Sounds in a Foreign Language Using Acoustic Measurements of the Vocal Tract as Feedback in Real Time.

    ERIC Educational Resources Information Center

    Dowd, Annette; Smith, John; Wolfe, Joe

    1998-01-01

    Measured the first two vowel-tract resonances of a sample of native-French speakers for the non-nasalized vowels of that language. Values measured for native speakers for a particular vowel were used as target parameters for subjects who used a visual display of an impedance spectrum of their own vocal tracts as real time feedback to realize the…

  20. Sex differences in the acoustic structure of vowel-like grunt vocalizations in baboons and their perceptual discrimination by baboon listeners

    NASA Astrophysics Data System (ADS)

    Rendall, Drew; Owren, Michael J.; Weerts, Elise; Hienz, Robert D.

    2004-01-01

    This study quantifies sex differences in the acoustic structure of vowel-like grunt vocalizations in baboons (Papio spp.) and tests the basic perceptual discriminability of these differences to baboon listeners. Acoustic analyses were performed on 1028 grunts recorded from 27 adult baboons (11 males and 16 females) in southern Africa, focusing specifically on the fundamental frequency (F0) and formant frequencies. The mean F0 and the mean frequencies of the first three formants were all significantly lower in males than they were in females, more dramatically so for F0. Experiments using standard psychophysical procedures subsequently tested the discriminability of adult male and adult female grunts. After learning to discriminate the grunt of one male from that of one female, five baboon subjects subsequently generalized this discrimination both to new call tokens from the same individuals and to grunts from novel males and females. These results are discussed in the context of both the possible vocal anatomical basis for sex differences in call structure and the potential perceptual mechanisms involved in their processing by listeners, particularly as these relate to analogous issues in human speech production and perception.

  1. Interspeaker Variability in Hard Palate Morphology and Vowel Production

    ERIC Educational Resources Information Center

    Lammert, Adam; Proctor, Michael; Narayanan, Shrikanth

    2013-01-01

    Purpose: Differences in vocal tract morphology have the potential to explain interspeaker variability in speech production. The potential acoustic impact of hard palate shape was examined in simulation, in addition to the interplay among morphology, articulation, and acoustics in real vowel production data. Method: High-front vowel production from…

  2. English vowel identification and vowel formant discrimination by native Mandarin Chinese- and native English-speaking listeners: The effect of vowel duration dependence.

    PubMed

    Mi, Lin; Tao, Sha; Wang, Wenjing; Dong, Qi; Guan, Jingjing; Liu, Chang

    2016-03-01

    The purpose of this study was to examine the relationship between English vowel identification and English vowel formant discrimination for native Mandarin Chinese- and native English-speaking listeners. The identification of 12 English vowels was measured with the duration cue preserved or removed. The thresholds of vowel formant discrimination on the F2 of two English vowels,/Λ/and/i/, were also estimated using an adaptive-tracking procedure. Native Mandarin Chinese-speaking listeners showed significantly higher thresholds of vowel formant discrimination and lower identification scores than native English-speaking listeners. The duration effect on English vowel identification was similar between native Mandarin Chinese- and native English-speaking listeners. Moreover, regardless of listeners' language background, vowel identification was significantly correlated with vowel formant discrimination for the listeners who were less dependent on duration cues, whereas the correlation between vowel identification and vowel formant discrimination was not significant for the listeners who were highly dependent on duration cues. This study revealed individual variability in using multiple acoustic cues to identify English vowels for both native and non-native listeners.

  3. Changes in Vowel Articulation with Subthalamic Nucleus Deep Brain Stimulation in Dysarthric Speakers with Parkinson's Disease

    PubMed Central

    Langlois, Mélanie; Prud'Homme, Michel; Cantin, Léo

    2014-01-01

    Purpose. To investigate changes in vowel articulation with the electrical deep brain stimulation (DBS) of the subthalamic nucleus (STN) in dysarthric speakers with Parkinson's disease (PD). Methods. Eight Quebec-French speakers diagnosed with idiopathic PD who had undergone STN DBS were evaluated ON-stimulation and OFF-stimulation (1 hour after DBS was turned off). Vowel articulation was compared ON-simulation versus OFF-stimulation using acoustic vowel space and formant centralization ratio, calculated with the first (F1) and second formant (F2) of the vowels /i/, /u/, and /a/. The impact of the preceding consonant context on articulation, which represents a measure of coarticulation, was also analyzed as a function of the stimulation state. Results. Maximum vowel articulation increased during ON-stimulation. Analyses also indicate that vowel articulation was modulated by the consonant context but this relationship did not change with STN DBS. Conclusions. Results suggest that STN DBS may improve articulation in dysarthric speakers with PD, in terms of range of movement. Optimization of the electrical parameters for each patient is important and may lead to improvement in speech fine motor control. However, the impact on overall speech intelligibility may still be small. Clinical considerations are discussed and new research avenues are suggested. PMID:25400977

  4. Articulatory Changes in Vowel Production following STN DBS and Levodopa Intake in Parkinson's Disease

    PubMed Central

    Martel Sauvageau, Vincent; Roy, Johanna-Pascale; Cantin, Léo; Prud'Homme, Michel; Langlois, Mélanie; Macoir, Joël

    2015-01-01

    Purpose. To investigate the impact of deep brain stimulation of the subthalamic nucleus (STN DBS) and levodopa intake on vowel articulation in dysarthric speakers with Parkinson's disease (PD). Methods. Vowel articulation was assessed in seven Quebec French speakers diagnosed with idiopathic PD who underwent STN DBS. Assessments were conducted on- and off-medication, first prior to surgery and then 1 year later. All recordings were made on-stimulation. Vowel articulation was measured using acoustic vowel space and formant centralization ratio. Results. Compared to the period before surgery, vowel articulation was reduced after surgery when patients were off-medication, while it was better on-medication. The impact of levodopa intake on vowel articulation changed with STN DBS: before surgery, levodopa impaired articulation, while it no longer had a negative effect after surgery. Conclusions. These results indicate that while STN DBS could lead to a direct deterioration in articulation, it may indirectly improve it by reducing the levodopa dose required to manage motor symptoms. These findings suggest that, with respect to speech production, STN DBS and levodopa intake cannot be investigated separately because the two are intrinsically linked. Along with motor symptoms, speech production should be considered when optimizing therapeutic management of patients with PD. PMID:26558134

  5. Native dialect influences second-language vowel perception: Peruvian versus Iberian Spanish learners of Dutch.

    PubMed

    Escudero, Paola; Williams, Daniel

    2012-05-01

    Peruvian Spanish (PS) and Iberian Spanish (IS) learners were tested on their ability to categorically discriminate and identify Dutch vowels. It was predicted that the acoustic differences between the vowel productions of the two dialects, which compare differently to Dutch vowels, would manifest in differential L2 perception for listeners of these two dialects. The results show that although PS learners had higher general L2 proficiency, IS learners were more accurate at discriminating all five contrasts and at identifying six of the L2 Dutch vowels. These findings confirm that acoustic differences in native vowel production lead to differential L2 vowel perception.

  6. Direct Mapping of Acoustics to Phonology: On the Lexical Encoding of Front Rounded Vowels in L1 English-L2 French Acquisition

    ERIC Educational Resources Information Center

    Darcy, Isabelle; Dekydtspotter, Laurent; Sprouse, Rex A.; Glover, Justin; Kaden, Christiane; McGuire, Michael; Scott, John H. G.

    2012-01-01

    It is well known that adult US-English-speaking learners of French experience difficulties acquiring high /y/-/u/ and mid /oe/-/[openo]/ front vs. back rounded vowel contrasts in French. This study examines the acquisition of these French vowel contrasts at two levels: phonetic categorization and lexical representations. An ABX categorization task…

  7. Neural Processing of Acoustic Duration and Phonological German Vowel Length: Time Courses of Evoked Fields in Response to Speech and Nonspeech Signals

    ERIC Educational Resources Information Center

    Tomaschek, Fabian; Truckenbrodt, Hubert; Hertrich, Ingo

    2013-01-01

    Recent experiments showed that the perception of vowel length by German listeners exhibits the characteristics of categorical perception. The present study sought to find the neural activity reflecting categorical vowel length and the short-long boundary by examining the processing of non-contrastive durations and categorical length using MEG.…

  8. Speech after Radial Forearm Free Flap Reconstruction of the Tongue: A Longitudinal Acoustic Study of Vowel and Diphthong Sounds

    ERIC Educational Resources Information Center

    Laaksonen, Juha-Pertti; Rieger, Jana; Happonen, Risto-Pekka; Harris, Jeffrey; Seikaly, Hadi

    2010-01-01

    The purpose of this study was to use acoustic analyses to describe speech outcomes over the course of 1 year after radial forearm free flap (RFFF) reconstruction of the tongue. Eighteen Canadian English-speaking females and males with reconstruction for oral cancer had speech samples recorded (pre-operative, and 1 month, 6 months, and 1 year…

  9. Relationships between objective acoustic indices and acoustic comfort evaluation in nonacoustic spaces

    NASA Astrophysics Data System (ADS)

    Kang, Jian

    2001-05-01

    Much attention has been paid to acoustic spaces such as concert halls and recording studios, whereas research on nonacoustic buildings/spaces has been rather limited, especially from the viewpoint of acoustic comfort. In this research a series of case studies has been carried out on this topic, considering various spaces including shopping mall atrium spaces, library reading rooms, football stadia, swimming spaces, churches, dining spaces, as well as urban open public spaces. The studies focus on the relationships between objective acoustic indices such as sound pressure level and reverberation time and perceptions of acoustic comfort. The results show that the acoustic atmosphere is an important consideration in such spaces and the evaluation of acoustic comfort may vary considerably even if the objective acoustic indices are the same. It is suggested that current guidelines and technical regulations are insufficient in terms of acoustic design of these spaces, and the relationships established from the case studies between objective and subjective aspects would be useful for developing further design guidelines. [Work supported partly by the British Academy.

  10. Documentation of the space station/aircraft acoustic apparatus

    NASA Technical Reports Server (NTRS)

    Clevenson, Sherman A.

    1987-01-01

    This paper documents the design and construction of the Space Station/Aircraft Acoustic Apparatus (SS/AAA). Its capabilities both as a space station acoustic simulator and as an aircraft acoustic simulator are described. Also indicated are the considerations which ultimately resulted in man-rating the SS/AAA. In addition, the results of noise surveys and reverberation time and absorption coefficient measurements are included.

  11. Space vehicle acoustics prediction improvement for payloads. [space shuttle

    NASA Technical Reports Server (NTRS)

    Dandridge, R. E.

    1979-01-01

    The modal analysis method was extensively modified for the prediction of space vehicle noise reduction in the shuttle payload enclosure, and this program was adapted to the IBM 360 computer. The predicted noise reduction levels for two test cases were compared with experimental results to determine the validity of the analytical model for predicting space vehicle payload noise environments in the 10 Hz one-third octave band regime. The prediction approach for the two test cases generally gave reasonable magnitudes and trends when compared with the measured noise reduction spectra. The discrepancies in the predictions could be corrected primarily by improved modeling of the vehicle structural walls and of the enclosed acoustic space to obtain a more accurate assessment of normal modes. Techniques for improving and expandng the noise prediction for a payload environment are also suggested.

  12. Acoustic emission technology for space applications

    SciTech Connect

    Friesel, M.A.; Lemon, D.K.; Skorpik, J.R.; Hutton, P.H.

    1989-05-01

    Clearly the structural and functional integrity of space station components is a primary requirement. The combinations of advanced materials, new designs, and an unusual environment increase the need for inservice monitoring to help assure component integrity. Continuous monitoring of the components using acoustic emission (AE) methods can provide early indication of structural or functional distress, thus allowing time to plan remedial action. The term ''AE'' refers to energy impulses propagated from a growing crack in a solid material or from a leak in a pressurized pipe or tube. In addition to detecting a crack or leak, AE methods can provide information on the location of the defect and an estimate of crack growth rate and leak rate. 8 figs.

  13. Perceptual Adaptation of Voice Gender Discrimination with Spectrally Shifted Vowels

    ERIC Educational Resources Information Center

    Li, Tianhao; Fu, Qian-Jie

    2011-01-01

    Purpose: To determine whether perceptual adaptation improves voice gender discrimination of spectrally shifted vowels and, if so, which acoustic cues contribute to the improvement. Method: Voice gender discrimination was measured for 10 normal-hearing subjects, during 5 days of adaptation to spectrally shifted vowels, produced by processing the…

  14. Influences of listeners' native and other dialects on cross-language vowel perception

    PubMed Central

    Williams, Daniel; Escudero, Paola

    2014-01-01

    This paper examines to what extent acoustic similarity between native and non-native vowels predicts non-native vowel perception and whether this process is influenced by listeners' native and other non-native dialects. Listeners with Northern and Southern British English dialects completed a perceptual assimilation task in which they categorized tokens of 15 Dutch vowels in terms of English vowel categories. While the cross-language acoustic similarity of Dutch vowels to English vowels largely predicted Southern listeners' perceptual assimilation patterns, this was not the case for Northern listeners, whose assimilation patterns resembled those of Southern listeners for all but three Dutch vowels. The cross-language acoustic similarity of Dutch vowels to Northern English vowels was re-examined by incorporating Southern English tokens, which resulted in considerable improvements in the predicting power of cross-language acoustic similarity. This suggests that Northern listeners' assimilation of Dutch vowels to English vowels was influenced by knowledge of both native Northern and non-native Southern English vowel categories. The implications of these findings for theories of non-native speech perception are discussed. PMID:25339921

  15. Production and perception of whispered vowels

    NASA Astrophysics Data System (ADS)

    Kiefte, Michael

    2005-09-01

    Information normally associated with pitch, such as intonation, can still be conveyed in whispered speech despite the absence of voicing. For example, it is possible to whisper the question ``You are going today?'' without any syntactic information to distinguish this sentence from a simple declarative. It has been shown that pitch change in whispered speech is correlated with the simultaneous raising or lowering of several formants [e.g., M. Kiefte, J. Acoust. Soc. Am. 116, 2546 (2004)]. However, spectral peak frequencies associated with formants have been identified as important correlates to vowel identity. Spectral peak frequencies may serve two roles in the perception of whispered speech: to indicate both vowel identity and intended pitch. Data will be presented to examine the relative importance of several acoustic properties including spectral peak frequencies and spectral shape parameters in both the production and perception of whispered vowels. Speakers were asked to phonate and whisper vowels at three different pitches across a range of roughly a musical fifth. It will be shown that relative spectral change is preserved within vowels across intended pitches in whispered speech. In addition, several models of vowel identification by listeners will be presented. [Work supported by SSHRC.

  16. Vowel Devoicing in Shanghai.

    ERIC Educational Resources Information Center

    Zee, Eric

    A phonetic study of vowel devoicing in the Shanghai dialect of Chinese explored the phonetic conditions under which the high, closed vowels and the apical vowel in Shanghai are most likely to become devoiced. The phonetic conditions may be segmental or suprasegmental. Segmentally, the study sought to determine whether a certain type of pre-vocalic…

  17. Articulatory characteristics of Hungarian ‘transparent’ vowels

    PubMed Central

    Benus, Stefan; Gafos, Adamantios I.

    2007-01-01

    Using a combination of magnetometry and ultrasound, we examined the articulatory characteristics of the so-called ‘transparent’ vowels [iː], [i], and [eː] in Hungarian vowel harmony. Phonologically, transparent vowels are front, but they can be followed by either front or back suffixes. However, a finer look reveals an underlying phonetic coherence in two respects. First, transparent vowels in back harmony contexts show a less advanced (more retracted) tongue body posture than phonemically identical vowels in front harmony contexts: e.g. [i] in buli-val is less advanced than [i] in bili-vel. Second, transparent vowels in monosyllabic stems selecting back suffixes are also less advanced than phonemically identical vowels in stems selecting front suffixes: e.g. [iː] in ír, taking back suffixes, compared to [iː] of hír, taking front suffixes, is less advanced when these stems are produced in bare form (no suffixes). We thus argue that the phonetic degree of tongue body horizontal position correlates with the phonological alternation in suffixes. A hypothesis that emerges from this work is that a plausible phonetic basis for transparency can be found in quantal characteristics of the relation between articulation and acoustics of transparent vowels. More broadly, the proposal is that the phonology of transparent vowels is better understood when their phonological patterning is studied together with their articulatory and acoustic characteristics. PMID:18389086

  18. Diversity of vowel systems

    NASA Astrophysics Data System (ADS)

    Maddieson, Ian

    2005-04-01

    Systems of vowels vary greatly across the world's languages while nonetheless conforming to certain general structural patterns. All languages have at least two qualitative distinctions between vowels based on the major parameters of height, backness and rounding, but probably none has more than 15 or so, and the modal number is 5. Generally these basic vowel qualities respect dispersion principles, but deviations can be considerable. When additional parameters, such as nasalization, length, phonation type and pharyngealization are included, the total number of vowel distinctions may easily exceed 40. These ``additive'' features never occur with a larger number of vowel qualities than those occurring in a ``plain'' series. Languages may differ markedly in the distributional patterns of their vowels as well as in their inventory. Some languages have different (usually reduced) vowel inventories in unstressed or other non-prominent positions; others constrain vowel sequences in (phonological) words through vowel harmony limitations. Co-occurrence patterns between vowels and consonants also vary greatly, as does the degree of coarticulation between vowels and neighboring segments. Learners must master all of these factors to speak an individual language fluently. Constraints that are universal or shared may be expected to facilitate this task.

  19. Vowel length in Farsi

    NASA Astrophysics Data System (ADS)

    Shademan, Shabnam

    2001-05-01

    This study tests whether Farsi vowels are contrastive with respective to length. Farsi has a six-vowel system with three lax vowels and three tense vowels. Both traditional grammarians and modern linguists believe that Farsi tense vowels are longer than lax vowels, and that there are no vowel pairs that contrast only in length. However, it has been suggested that Farsi exhibits compensatory lengthening, which is triggered by the deletion of glottal consonants in coda position in informal speech (Darzi, 1991). As a result, minimal pairs such as [tar] and [tarh] should contrast only with respect to vowel length. A corpus of 90 words of the form CVC, CVCG, CVGC, and CVCC (where V=a vowel and G=a glottal consonant) was recorded, and durations of vowels in different contexts were measured and compared. Preliminary results show that lax vowel durations fall into three groups with CVCC longer than CVCG/CVGC, and the latter longer than CVC. It remains to be seen whether CVCG/CVGC words show compensatory lengthening when the glottal consonant is deleted.

  20. Adult Second Language Learning of Spanish Vowels

    ERIC Educational Resources Information Center

    Cobb, Katherine; Simonet, Miquel

    2015-01-01

    The present study reports on the findings of a cross-sectional acoustic study of the production of Spanish vowels by three different groups of speakers: 1) native Spanish speakers; 2) native English intermediate learners of Spanish; and 3) native English advanced learners of Spanish. In particular, we examined the production of the five Spanish…

  1. Two Notes on Kinande Vowel Harmony

    ERIC Educational Resources Information Center

    Kenstowicz, Michael J.

    2009-01-01

    This paper documents the acoustic reflexes of ATR harmony in Kinande followed by an analysis of the dominance reversal found in class 5 nominals. The principal findings are that the ATR harmony is reliably reflected in a lowering of the first formant. Depending on the vowel, ATR harmony also affects the second formant. The directional asymmetry…

  2. American and Swedish children's acquisition of vowel duration: Effects of vowel identity and final stop voicing

    NASA Astrophysics Data System (ADS)

    Buder, Eugene H.; Stoel-Gammon, Carol

    2002-04-01

    Vowel durations typically vary according to both intrinsic (segment-specific) and extrinsic (contextual) specifications. It can be argued that such variations are due to both predisposition and cognitive learning. The present report utilizes acoustic phonetic measurements from Swedish and American children aged 24 and 30 months to investigate the hypothesis that default behaviors may precede language-specific learning effects. The predicted pattern is the presence of final consonant voicing effects in both languages as a default, and subsequent learning of intrinsic effects most notably in the Swedish children. The data, from 443 monosyllabic tokens containing high-front vowels and final stop consonants, are analyzed in statistical frameworks at group and individual levels. The results confirm that Swedish children show an early tendency to vary vowel durations according to final consonant voicing, followed only six months later by a stage at which the intrinsic influence of vowel identity grows relatively more robust. Measures of vowel formant structure from selected 30-month-old children also revealed a tendency for children of this age to focus on particular acoustic contrasts. In conclusion, the results indicate that early acquisition of vowel specifications involves an interaction between language-specific features and articulatory predispositions associated with phonetic context.

  3. Effects of Intensive Voice Treatment (the Lee Silverman Voice Treatment [LSVT]) on Vowel Articulation in Dysarthric Individuals with Idiopathic Parkinson Disease: Acoustic and Perceptual Findings

    ERIC Educational Resources Information Center

    Sapir, Shimon; Spielman, Jennifer L.; Ramig, Lorraine O.; Story, Brad H.; Fox, Cynthia

    2007-01-01

    Purpose: To evaluate the effects of intensive voice treatment targeting vocal loudness (the Lee Silverman Voice Treatment [LSVT]) on vowel articulation in dysarthric individuals with idiopathic Parkinson's disease (PD). Method: A group of individuals with PD receiving LSVT (n = 14) was compared to a group of individuals with PD not receiving LSVT…

  4. Modeling Consonant-Vowel Coarticulation for Articulatory Speech Synthesis

    PubMed Central

    Birkholz, Peter

    2013-01-01

    A central challenge for articulatory speech synthesis is the simulation of realistic articulatory movements, which is critical for the generation of highly natural and intelligible speech. This includes modeling coarticulation, i.e., the context-dependent variation of the articulatory and acoustic realization of phonemes, especially of consonants. Here we propose a method to simulate the context-sensitive articulation of consonants in consonant-vowel syllables. To achieve this, the vocal tract target shape of a consonant in the context of a given vowel is derived as the weighted average of three measured and acoustically-optimized reference vocal tract shapes for that consonant in the context of the corner vowels /a/, /i/, and /u/. The weights are determined by mapping the target shape of the given context vowel into the vowel subspace spanned by the corner vowels. The model was applied for the synthesis of consonant-vowel syllables with the consonants /b/, /d/, /g/, /l/, /r/, /m/, /n/ in all combinations with the eight long German vowels. In a perception test, the mean recognition rate for the consonants in the isolated syllables was 82.4%. This demonstrates the potential of the approach for highly intelligible articulatory speech synthesis. PMID:23613734

  5. Acoustic emissions applications on the NASA Space Station

    SciTech Connect

    Friesel, M.A.; Dawson, J.F.; Kurtz, R.J.; Barga, R.S.; Hutton, P.H.; Lemon, D.K.

    1991-08-01

    Acoustic emission is being investigated as a way to continuously monitor the space station Freedom for damage caused by space debris impact and seal failure. Experiments run to date focused on detecting and locating simulated and real impacts and leakage. These were performed both in the laboratory on a section of material similar to a space station shell panel and also on the full-scale common module prototype at Boeing's Huntsville facility. A neural network approach supplemented standard acoustic emission detection and analysis techniques. 4 refs., 5 figs., 1 tab.

  6. Acoustic levitation for high temperature containerless processing in space

    NASA Technical Reports Server (NTRS)

    Rey, C. A.; Sisler, R.; Merkley, D. R.; Danley, T. J.

    1990-01-01

    New facilities for high-temperature containerless processing in space are described, including the acoustic levitation furnace (ALF), the high-temperature acoustic levitator (HAL), and the high-pressure acoustic levitator (HPAL). In the current ALF development, the maximum temperature capabilities of the levitation furnaces are 1750 C, and in the HAL development with a cold wall furnace they will exceed 2000-2500 C. The HPAL demonstrated feasibility of precursor space flight experiments on the ground in a 1 g pressurized-gas environment. Testing of lower density materials up to 1300 C has also been accomplished. It is suggested that advances in acoustic levitation techniques will result in the production of new materials such as ceramics, alloys, and optical and electronic materials.

  7. Auditory, visual, and auditory-visual perception of vowels by hearing-impaired children.

    PubMed

    Hack, Z C; Erber, N P

    1982-03-01

    The vowels (foreign letters in text) were presented through auditory, visual, and combined auditory-visual modalities to hearing-impaired children having good, intermediate, and poor auditory work-recognition skills. When they received acoustic information only, children with good word-recognition skills confused neighboring vowels (i.e., those having similar formant frequencies). Children with intermediate work-recognition skills demonstrated this same difficulty and confused front and back vowels. Children with poor word-recognition skills identified vowels mainly on the basis of temporal and intensity cues. Through lipreading alone, all three groups distinguished spread from rounded vowels but could not reliably identify vowels within the categories. The first two groups exhibited only moderate difficulty in identifying vowels audiovisually. The third group, although showing a small amount of improvement over lipreading alone, still experienced difficulty in identifying vowels through combined auditory and visual modes.

  8. Threshold differences for interaural time delays carried by double vowels

    NASA Astrophysics Data System (ADS)

    Akeroyd, Michael A.

    2003-10-01

    Experimental measurements were made of threshold interaural time differences (ITDs) for a ``target'' vowel presented simultaneously with a fixed-ITD ``distracter'' vowel. Three double-vowel pairs were used, comprising an ``er'' (/schwa/) together with either an ``ai,'' ``ar,'' or ``oo'' (respectively, /ee/, /openo/, and /you/). Threshold ITDs were found to be larger for the target vowel when it was part of a double-vowel pair than in control conditions in which it was presented alone. The effect size depended upon the choice of target vowel and distracter vowel, the level of the target relative to the distracter, and whether the two vowels had the same or different fundamental frequencies. The experiment was analyzed using a multichannel modification of Heller and Trahiotis' [J. Acoust. Soc. Am. 99, 3632-3637 (1996)] model, which used a weighted combination of the detectabilities of the ITD of the target and the distracter. It gave predictions consistent with the observed effects of level and with some of the effects of the choice of target vowel, but it could not describe the effect of the target-distracter differences in fundamental frequency. It was found that a single-channel version of the model, in which the chosen channel was allowed to depend upon fundamental frequency (which could be derived using a monaural autocorrelation model) did give a set of predictions in qualitative accord with the data.

  9. The Shift in Infant Preferences for Vowel Duration and Pitch Contour between 6 and 10 Months of Age

    ERIC Educational Resources Information Center

    Kitamura, Christine; Notley, Anna

    2009-01-01

    This study investigates the influence of the acoustic properties of vowels on 6- and 10-month-old infants' speech preferences. The shape of the contour (bell or monotonic) and the duration (normal or stretched) of vowels were manipulated in words containing the vowels /i/ and /u/, and presented to infants using a two-choice preference procedure.…

  10. Vowel Intelligibility in Children with and without Dysarthria: An Exploratory Study

    ERIC Educational Resources Information Center

    Levy, Erika S.; Leone, Dorothy; Moya-Gale, Gemma; Hsu, Sih-Chiao; Chen, Wenli; Ramig, Lorraine O.

    2016-01-01

    Children with dysarthria due to cerebral palsy (CP) present with decreased vowel space area and reduced word intelligibility. Although a robust relationship exists between vowel space and word intelligibility, little is known about the intelligibility of vowels in this population. This exploratory study investigated the intelligibility of American…

  11. Space manufacturing of surface acoustic wave devices, appendix D

    NASA Technical Reports Server (NTRS)

    Sardella, G.

    1973-01-01

    Space manufacturing of transducers in a vibration free environment is discussed. Fabrication of the masks, and possible manufacturing of the surface acoustic wave components aboard a space laboratory would avoid the inherent ground vibrations and the frequency limitation imposed by a seismic isolator pad. The manufacturing vibration requirements are identified. The concepts of space manufacturing are analyzed. A development program for manufacturing transducers is recommended.

  12. International Space Station Crew Quarters Ventilation and Acoustic Design Implementation

    NASA Technical Reports Server (NTRS)

    Broyan, James L., Jr.; Cady, Scott M; Welsh, David A.

    2010-01-01

    The International Space Station (ISS) United States Operational Segment has four permanent rack sized ISS Crew Quarters (CQs) providing a private crew member space. The CQs use Node 2 cabin air for ventilation/thermal cooling, as opposed to conditioned ducted air-from the ISS Common Cabin Air Assembly (CCAA) or the ISS fluid cooling loop. Consequently, CQ can only increase the air flow rate to reduce the temperature delta between the cabin and the CQ interior. However, increasing airflow causes increased acoustic noise so efficient airflow distribution is an important design parameter. The CQ utilized a two fan push-pull configuration to ensure fresh air at the crew member's head position and reduce acoustic exposure. The CQ ventilation ducts are conduits to the louder Node 2 cabin aisle way which required significant acoustic mitigation controls. The CQ interior needs to be below noise criteria curve 40 (NC-40). The design implementation of the CQ ventilation system and acoustic mitigation are very inter-related and require consideration of crew comfort balanced with use of interior habitable volume, accommodation of fan failures, and possible crew uses that impact ventilation and acoustic performance. Each CQ required 13% of its total volume and approximately 6% of its total mass to reduce acoustic noise. This paper illustrates the types of model analysis, assumptions, vehicle interactions, and trade-offs required for CQ ventilation and acoustics. Additionally, on-orbit ventilation system performance and initial crew feedback is presented. This approach is applicable to any private enclosed space that the crew will occupy.

  13. The Vietnamese Vowel System

    ERIC Educational Resources Information Center

    Emerich, Giang Huong

    2012-01-01

    In this dissertation, I provide a new analysis of the Vietnamese vowel system as a system with fourteen monophthongs and nineteen diphthongs based on phonetic and phonological data. I propose that these Vietnamese contour vowels - /ie/, /[turned m]?/ and /uo/-should be grouped with these eleven monophthongs /i e epsilon a [turned a] ? ? [turned m]…

  14. Producing American English Vowels during Vocal Tract Growth: A Perceptual Categorization Study of Synthesized Vowels

    ERIC Educational Resources Information Center

    Menard, Lucie; Davis, Barbara L.; Boe, Louis-Jean; Roy, Johanna-Pascale

    2009-01-01

    Purpose: To consider interactions of vocal tract change with growth and perceived output patterns across development, the influence of nonuniform vocal tract growth on the ability to reach acoustic-perceptual targets for English vowels was studied. Method: Thirty-seven American English speakers participated in a perceptual categorization…

  15. Space Launch System Begins Acoustic Testing

    NASA Video Gallery

    Engineers at NASA's Marshall Space Flight Center in Huntsville, Ala., have assembled a collection of thrusters to stand in for the various propulsion elements in a scale model version of NASA’s S...

  16. Vowel Formant Values in Hearing and Hearing-Impaired Children: A Discriminant Analysis

    ERIC Educational Resources Information Center

    Ozbic, Martina; Kogovsek, Damjana

    2010-01-01

    Hearing-impaired speakers show changes in vowel production and formant pitch and variability, as well as more cases of overlapping between vowels and more restricted formant space, than hearing speakers; consequently their speech is less intelligible. The purposes of this paper were to determine the differences in vowel formant values between 32…

  17. Acoustic positioning for space processing experiments

    NASA Technical Reports Server (NTRS)

    Whymark, R. R.

    1974-01-01

    An acoustic positioning system is described that is adaptable to a range of processing chambers and furnace systems. Operation at temperatures exceeding 1000 C is demonstrated in experiments involving the levitation of liquid and solid glass materials up to several ounces in weight. The system consists of a single source of sound that is beamed at a reflecting surface placed a distance away. Stable levitation is achieved at a succession of discrete energy minima contained throughout the volume between the reflector and the sound source. Several specimens can be handled at one time. Metal discs up to 3 inches in diameter can be levitated, solid spheres of dense material up to 0.75 inches diameter, and liquids can be freely suspended in l-g in the form of near-spherical droplets up to 0.25 inch diameter, or flattened liquid discs up to 0.6 inches diameter. Larger specimens may be handled by increasing the size of the sound source or by reducing the sound frequency.

  18. Vowel constrictions are recoverable from formants

    PubMed Central

    Iskarous, Khalil

    2010-01-01

    The area function of the vocal tract in all of its spatial detail is not directly computable from the speech signal. But is partial, yet phonetically distinctive, information about articulation recoverable from the acoustic signal that arrives at the listener’s ear? The answer to this question is important for phonetics, because various theories of speech perception predict different answers. Some theories assume that recovery of articulatory information must be possible, while others assume that it is impossible. However, neither type of theory provides firm evidence showing that distinctive articulatory information is or is not extractable from the acoustic signal. The present study focuses on vowel gestures and examines whether linguistically significant information, such as the constriction location, constriction degree, and rounding, is contained in the speech signal, and whether such information is recoverable from formant parameters. Perturbation theory and linear prediction were combined, in a manner similar to that in Mokhtari (1998) [Mokhtari, P. (1998). An acoustic-phonetic and articulatory study of speech-speaker dichotomy. Doctoral dissertation, University of New South Wales], to assess the accuracy of recovery of information about vowel constrictions. Distinctive constriction information estimated from the speech signal for ten American English vowels were compared to the constriction information derived from simultaneously collected X-ray microbeam articulatory data for 39 speakers [Westbury (1994). Xray microbeam speech production database user’s handbook. University of Wisconsin, Madison, WI]. The recovery of distinctive articulatory information relies on a novel technique that uses formant frequencies and amplitudes, and does not depend on a principal components analysis of the articulatory data, as do most other inversion techniques. These results provide evidence that distinctive articulatory information for vowels can be recovered from the

  19. Variability in Vowel Production within and between Days

    PubMed Central

    Heald, Shannon L. M.; Nusbaum, Howard C.

    2015-01-01

    Although the acoustic variability of speech is often described as a problem for phonetic recognition, there is little research examining acoustic-phonetic variability over time. We measured naturally occurring acoustic variability in speech production at nine specific time points (three per day over three days) to examine daily change in production as well as change across days for citation-form vowels. Productions of seven different vowels (/EE/, /IH/, /AH/, /UH/, /AE/, /OO/, /EH/) were recorded at 9AM, 3PM and 9PM over the course of each testing day on three different days, every other day, over a span of five days. Results indicate significant systematic change in F1 and F0 values over the course of a day for each of the seven vowels recorded, whereas F2 and F3 remained stable. Despite this systematic change within a day, however, talkers did not show significant changes in F0, F1, F2, and F3 between days, demonstrating that speakers are capable of producing vowels with great reliability over days without any extrinsic feedback besides their own auditory monitoring. The data show that in spite of substantial day-to-day variability in the specific listening and speaking experiences of these participants and thus exposure to different acoustic tokens of speech, there is a high degree of internal precision and consistency for the production of citation form vowels. PMID:26331478

  20. The vowel systems of Quichua-Spanish bilinguals. Age of acquisition effects on the mutual influence of the first and second languages.

    PubMed

    Guion, Susan G

    2003-01-01

    This study investigates vowel productions of 20 Quichua-Spanish bilinguals, differing in age of Spanish acquisition, and 5 monolingual Spanish speakers. While the vowel systems of simultaneous, early, and some mid bilinguals all showed significant plasticity, there were important differences in the kind, as well as the extent, of this adaptability. Simultaneous bilinguals differed from early bilinguals in that they were able to partition the vowel space in a more fine-grained way to accommodate the vowels of their two languages. Early and some mid bilinguals acquired Spanish vowels, whereas late bilinguals did not. It was also found that acquiring Spanish vowels could affect the production of native Quichua vowels. The Quichua vowels were produced higher by bilinguals who had acquired Spanish vowels than those who had not. It is proposed that this vowel reorganization serves to enhance the perceptual distinctiveness between the vowels of the combined first- and second-language system. PMID:12853715

  1. Retroactive Streaming Fails to Improve Concurrent Vowel Identification

    PubMed Central

    2015-01-01

    The sequential organization of sound over time can interact with the concurrent organization of sounds across frequency. Previous studies using simple acoustic stimuli have suggested that sequential streaming cues can retroactively affect the perceptual organization of sounds that have already occurred. It is unknown whether such effects generalize to the perception of speech sounds. Listeners’ ability to identify two simultaneously presented vowels was measured in the following conditions: no context, a preceding context stream (precursors), and a following context stream (postcursors). The context stream was comprised of brief repetitions of one of the two vowels, and the primary measure of performance was listeners’ ability to identify the other vowel. Results in the precursor condition showed a significant advantage for the identification of the second vowel compared to the no-context condition, suggesting that sequential grouping mechanisms aided the segregation of the concurrent vowels, in agreement with previous work. However, performance in the postcursor condition was significantly worse compared to the no-context condition, providing no evidence for an effect of stream segregation, and suggesting a possible interference effect. Two additional experiments involving inharmonic (jittered) vowels were performed to provide additional cues to aid retroactive stream segregation; however, neither manipulation enabled listeners to improve their identification of the target vowel. Taken together with earlier studies, the results suggest that retroactive streaming may require large spectral differences between concurrent sources and thus may not provide a robust segregation cue for natural broadband sounds such as speech. PMID:26451598

  2. Retroactive Streaming Fails to Improve Concurrent Vowel Identification.

    PubMed

    Brandewie, Eugene J; Oxenham, Andrew J

    2015-01-01

    The sequential organization of sound over time can interact with the concurrent organization of sounds across frequency. Previous studies using simple acoustic stimuli have suggested that sequential streaming cues can retroactively affect the perceptual organization of sounds that have already occurred. It is unknown whether such effects generalize to the perception of speech sounds. Listeners' ability to identify two simultaneously presented vowels was measured in the following conditions: no context, a preceding context stream (precursors), and a following context stream (postcursors). The context stream was comprised of brief repetitions of one of the two vowels, and the primary measure of performance was listeners' ability to identify the other vowel. Results in the precursor condition showed a significant advantage for the identification of the second vowel compared to the no-context condition, suggesting that sequential grouping mechanisms aided the segregation of the concurrent vowels, in agreement with previous work. However, performance in the postcursor condition was significantly worse compared to the no-context condition, providing no evidence for an effect of stream segregation, and suggesting a possible interference effect. Two additional experiments involving inharmonic (jittered) vowels were performed to provide additional cues to aid retroactive stream segregation; however, neither manipulation enabled listeners to improve their identification of the target vowel. Taken together with earlier studies, the results suggest that retroactive streaming may require large spectral differences between concurrent sources and thus may not provide a robust segregation cue for natural broadband sounds such as speech. PMID:26451598

  3. Toward a Systematic Evaluation of Vowel Target Events across Speech Tasks

    ERIC Educational Resources Information Center

    Kuo, Christina

    2011-01-01

    The core objective of this study was to examine whether acoustic variability of vowel production in American English, across speaking tasks, is systematic. Ten male speakers who spoke a relatively homogeneous Wisconsin dialect produced eight monophthong vowels (in hVd and CVC contexts) in four speaking tasks, including clear-speech, citation form,…

  4. Acoustic Emission Detection of Impact Damage on Space Shuttle Structures

    NASA Technical Reports Server (NTRS)

    Prosser, William H.; Gorman, Michael R.; Madaras, Eric I.

    2004-01-01

    The loss of the Space Shuttle Columbia as a result of impact damage from foam debris during ascent has led NASA to investigate the feasibility of on-board impact detection technologies. AE sensing has been utilized to monitor a wide variety of impact conditions on Space Shuttle components ranging from insulating foam and ablator materials, and ice at ascent velocities to simulated hypervelocity micrometeoroid and orbital debris impacts. Impact testing has been performed on both reinforced carbon composite leading edge materials as well as Shuttle tile materials on representative aluminum wing structures. Results of these impact tests will be presented with a focus on the acoustic emission sensor responses to these impact conditions. These tests have demonstrated the potential of employing an on-board Shuttle impact detection system. We will describe the present plans for implementation of an initial, very low frequency acoustic impact sensing system using pre-existing flight qualified hardware. The details of an accompanying flight measurement system to assess the Shuttle s acoustic background noise environment as a function of frequency will be described. The background noise assessment is being performed to optimize the frequency range of sensing for a planned future upgrade to the initial impact sensing system.

  5. The phonological function of vowels is maintained at fundamental frequencies up to 880 Hz.

    PubMed

    Friedrichs, Daniel; Maurer, Dieter; Dellwo, Volker

    2015-07-01

    In a between-subject perception task, listeners either identified full words or vowels isolated from these words at F0s between 220 and 880 Hz. They received two written words as response options (minimal pair with the stimulus vowel in contrastive position). Listeners' sensitivity (A') was extremely high in both conditions at all F0s, showing that the phonological function of vowels can also be maintained at high F0s. This indicates that vowel sounds may carry strong acoustic cues departing from common formant frequencies at high F0s and that listeners do not rely on consonantal context phenomena for their identification performance. PMID:26233058

  6. Perception of sinewave vowels

    PubMed Central

    Hillenbrand, James M.; Clark, Michael J.; Baer, Carter A.

    2011-01-01

    There is a significant body of research examining the intelligibility of sinusoidal replicas of natural speech. Discussion has followed about what the sinewave speech phenomenon might imply about the mechanisms underlying phonetic recognition. However, most of this work has been conducted using sentence material, making it unclear what the contributions are of listeners’ use of linguistic constraints versus lower level phonetic mechanisms. This study was designed to measure vowel intelligibility using sinusoidal replicas of naturally spoken vowels. The sinusoidal signals were modeled after 300 ∕hVd∕ syllables spoken by men, women, and children. Students enrolled in an introductory phonetics course served as listeners. Recognition rates for the sinusoidal vowels averaged 55%, which is much lower than the ∼95% intelligibility of the original signals. Attempts to improve performance using three different training methods met with modest success, with post-training recognition rates rising by ∼5–11 percentage points. Follow-up work showed that more extensive training produced further improvements, with performance leveling off at ∼73%–74%. Finally, modeling work showed that a fairly simple pattern-matching algorithm trained on naturally spoken vowels classified sinewave vowels with 78.3% accuracy, showing that the sinewave speech phenomenon does not necessarily rule out template matching as a mechanism underlying phonetic recognition. PMID:21682420

  7. Reading skills and the discrimination of English vowel contrasts by bilingual Spanish/English-speaking children: Is there a correlation?

    NASA Astrophysics Data System (ADS)

    Levey, Sandra

    2005-04-01

    This study examined the discrimination of English vowel contrasts in real and novel word-pairs by 21 children: 11 bilingual Spanish/English- and 10 monolingual English-speaking children, 8-12 years of age (M=10; 6; Mdn=10; 4). The goal was to determine if children with poor reading skills had difficulty with discrimination, an essential factor in reading abilities. A categorial discrimination task was used in an ABX discrimination paradigm: A (the first word in the sequence) and B (the second word in the sequence) were different stimuli, and X (the third word in the sequence) was identical to either A or to B. Stimuli were produced by one of three different speakers. Seventy-two monosyllabic words were presented: 36 real English and 36 novel words. Vowels were those absent from the inventory of Spanish vowels. Discrimination accuracy for the English-speaking children with good reading skills was significantly greater than for the bilingual-speaking children with good or poor reading skills. Early age of acquisition and greater percentage of time devoted to communication in English played the greatest role in bilingual children's discrimination and reading skills. The adjacency of vowels in the F1-F2 acoustic space presented the greatest difficulty.

  8. Acoustics in the Worship Space: Goals and Strategies

    NASA Astrophysics Data System (ADS)

    Crist, Ernest Vincent, III

    The act of corporate worship, though encompassing nearly all the senses, is primarily aural. Since the transmission process is aural, the experiences must be heard with sufficient volume and quality in order to be effective. The character and magnitude of the sound-producing elements are significant in the process, but it is the nature of the enclosing space that is even more crucial. Every building will, by its size, shape, and materials, act upon all sounds produced, affecting not only the amount, but also the quality of sound reaching the listeners' ears. The purpose of this project was to determine appropriate acoustical goals for worship spaces, and to propose design strategies to achieve these goals. Acoustic goals were determined by examination of the results of previously conducted subjective preference studies, computer plotting of virtual sources, and experimentation in actual spaces. Determinations also take into account the sensory inhibition factors in the processing of sound by the human auditory system. Design strategies incorporate aspects of placement of performing forces, the geometry of the enclosing space, materials and furnishings, and noise control.

  9. Formant Centralization Ratio: A Proposal for a New Acoustic Measure of Dysarthric Speech

    ERIC Educational Resources Information Center

    Sapir, Shimon; Ramig, Lorraine O.; Spielman, Jennifer L.; Fox, Cynthia

    2010-01-01

    Purpose: The vowel space area (VSA) has been used as an acoustic metric of dysarthric speech, but with varying degrees of success. In this study, the authors aimed to test an alternative metric to the VSA--the "formant centralization ratio" (FCR), which is hypothesized to more effectively differentiate dysarthric from healthy speech and register…

  10. Virtual Acoustic Space: Space perception for the blind

    NASA Astrophysics Data System (ADS)

    Rodríguez-Ramos, Luis F.

    2011-06-01

    This R&D project implements a new way of perception of the three-dimensional surrounding space, based exclusively in sounds and thus especially useful for the blind. The innate capability of locating sounds, the externalization of sounds played with headphones and the machine capture of the 3D environment are the technological pillars used for this purpose. They are analysed and a summary of their main requirements are presented. A number of laboratory facilities and portable prototypes are described, together with their main characteristics.

  11. Target spectral, dynamic spectral, and duration cues in infant perception of German vowels.

    PubMed

    Bohn, O S; Polka, L

    2001-07-01

    Previous studies of vowel perception have shown that adult speakers of American English and of North German identify native vowels by exploiting at least three types of acoustic information contained in consonant-vowel-consonant (CVC) syllables: target spectral information reflecting the articulatory target of the vowel, dynamic spectral information reflecting CV- and -VC coarticulation, and duration information. The present study examined the contribution of each of these three types of information to vowel perception in prelingual infants and adults using a discrimination task. Experiment 1 examined German adults' discrimination of four German vowel contrasts (see text), originally produced in /dVt/ syllables, in eight experimental conditions in which the type of vowel information was manipulated. Experiment 2 examined German-learning infants' discrimination of the same vowel contrasts using a comparable procedure. The results show that German adults and German-learning infants appear able to use either dynamic spectral information or target spectral information to discriminate contrasting vowels. With respect to duration information, the removal of this cue selectively affected the discriminability of two of the vowel contrasts for adults. However, for infants, removal of contrastive duration information had a larger effect on the discrimination of all contrasts tested. PMID:11508975

  12. Indexical properties influence time-varying amplitude and fundamental frequency contributions of vowels to sentence intelligibility

    PubMed Central

    Fogerty, Daniel

    2015-01-01

    The present study investigated how non-linguistic, indexical information about talker identity interacts with contributions to sentence intelligibility by the time-varying amplitude (temporal envelope) and fundamental frequency (F0). Young normal-hearing adults listened to sentences that preserved the original consonants but replaced the vowels with a single vowel production. This replacement vowel selectively preserved amplitude or F0 cues of the original vowel, but replaced cues to phonetic identity. Original vowel duration was always preserved. Three experiments investigated indexical contributions by replacing vowels with productions from the same or different talker, or by acoustically morphing the original vowel. These stimulus conditions investigated how vowel suprasegmental and indexical properties interact and contribute to intelligibility independently from phonetic information. Results demonstrated that indexical properties influence the relative contribution of suprasegmental properties to sentence intelligibility. F0 variations are particularly important in the presence of conflicting indexical information. Temporal envelope modulations significantly improve sentence intelligibility, but are enhanced when either indexical or F0 cues are available. These findings suggest that F0 and other indexical cues may facilitate perceptually grouping suprasegmental properties of vowels with the remainder of the sentence. Temporal envelope modulations of vowels may contribute to intelligibility once they are successfully integrated with the preserved signal. PMID:26543276

  13. Underwater acoustic source localization using closely spaced hydrophone pairs

    NASA Astrophysics Data System (ADS)

    Sim, Min Seop; Choi, Bok-Kyoung; Kim, Byoung-Nam; Lee, Kyun Kyung

    2016-07-01

    Underwater sound source position is determined using a line array. However, performance degradation occurs owing to a multipath environment, which generates incoherent signals. In this paper, a hydrophone array is proposed for underwater source position estimation robust to a multipath environment. The array is composed of three pairs of sensors placed on the same line. The source position is estimated by performing generalized cross-correlation (GCC). The proposed system is not affected by a multipath time delay because of the close distance between closely spaced sensors. The validity of the array is confirmed by simulation using acoustic signals synthesized by eigenrays.

  14. Acoustics

    NASA Technical Reports Server (NTRS)

    Goodman, Jerry R.; Grosveld, Ferdinand

    2007-01-01

    The acoustics environment in space operations is important to maintain at manageable levels so that the crewperson can remain safe, functional, effective, and reasonably comfortable. High acoustic levels can produce temporary or permanent hearing loss, or cause other physiological symptoms such as auditory pain, headaches, discomfort, strain in the vocal cords, or fatigue. Noise is defined as undesirable sound. Excessive noise may result in psychological effects such as irritability, inability to concentrate, decrease in productivity, annoyance, errors in judgment, and distraction. A noisy environment can also result in the inability to sleep, or sleep well. Elevated noise levels can affect the ability to communicate, understand what is being said, hear what is going on in the environment, degrade crew performance and operations, and create habitability concerns. Superfluous noise emissions can also create the inability to hear alarms or other important auditory cues such as an equipment malfunctioning. Recent space flight experience, evaluations of the requirements in crew habitable areas, and lessons learned (Goodman 2003; Allen and Goodman 2003; Pilkinton 2003; Grosveld et al. 2003) show the importance of maintaining an acceptable acoustics environment. This is best accomplished by having a high-quality set of limits/requirements early in the program, the "designing in" of acoustics in the development of hardware and systems, and by monitoring, testing and verifying the levels to ensure that they are acceptable.

  15. Infant directed speech in natural interaction--Norwegian vowel quantity and quality.

    PubMed

    Englund, Kjellrun T; Behne, Dawn M

    2005-05-01

    An interactive face-to-face setting is used to study natural infant directed speech (IDS) compared to adult directed speech (ADS). With distinctive vowel quantity and vowel quality, Norwegian IDS was used in a natural quasi-experimental design. Six Norwegian mothers were recorded over a period of 6 months alone with their infants and in an adult conversation. Vowel duration and spectral attributes of the vowels /a:/, /i:/ and /u:/, and their short counterparts /a/ /i/ and /u/ were analysed. Repeated measures analyses show that effects of vowel quantity did not differ between ADS and IDS, and for back vowel qualities, the vowel space was shifted upwards in IDS compared to ADS suggesting that fronted articulations in natural IDS may visually enhance speech to infants.

  16. The Effect of Stress and Speech Rate on Vowel Coarticulation in Catalan Vowel-Consonant-Vowel Sequences

    ERIC Educational Resources Information Center

    Recasens, Daniel

    2015-01-01

    Purpose: The goal of this study was to ascertain the effect of changes in stress and speech rate on vowel coarticulation in vowel-consonant-vowel sequences. Method: Data on second formant coarticulatory effects as a function of changing /i/ versus /a/ were collected for five Catalan speakers' productions of vowel-consonant-vowel sequences with the…

  17. Dynamic Articulation of Vowels.

    ERIC Educational Resources Information Center

    Morgan, Willie B.

    1979-01-01

    A series of exercises and a theory of vowel descriptions can help minimize speakers' problems of excessive tension, awareness of tongue height, and tongue retraction. Eight exercises to provide Forward Facial Stretch neutralize tensions in the face and vocal resonator and their effect on the voice. Three experiments in which sounds are repeated…

  18. Resolution of lateral acoustic space assessed by electroencephalography and psychoacoustics

    PubMed Central

    Bennemann, Jan; Freigang, Claudia; Schröger, Erich; Rübsamen, Rudolf; Richter, Nicole

    2013-01-01

    The encoding of auditory spatial acuity (measured as the precision to distinguish between two spatially distinct stimuli) by neural circuits in both auditory cortices is a matter of ongoing research. Here, the event-related potential (ERP) mismatch negativity (MMN), a sensitive indicator of preattentive auditory change detection, was used to tap into the underlying mechanism of cortical representation of auditory spatial information. We characterized the MMN response affected by the degree of spatial deviance in lateral acoustic space using a passive oddball paradigm. Two stimulation conditions (SCs)—specifically focusing on the investigation of the mid- and far-lateral acoustic space—were considered: (1) 65° left standard position with deviant positions at 70, 75, and 80°; and (2) 95° left standard position with deviant positions at 90, 85, and 80°. Additionally, behavioral data on the minimum audible angle (MAA) were acquired for the respective standard positions (65, 95° left) to quantify spatial discrimination in separating distinct sound sources. The two measurements disclosed the linkage between the (preattentive) MMN response and the (attentive) behavioral threshold. At 65° spatial deviations as small as 5° reliably elicited MMNs. Thereby, the MMN amplitudes monotonously increased as a function of spatial deviation. At 95°, spatial deviations of 15° were necessary to elicit a valid MMN. The behavioral data, however, yielded no difference in mean MAA thresholds for position 65 and 95°. The different effects of laterality on MMN responses and MAA thresholds suggest a role of spatial selective attention mechanisms particularly relevant in active discrimination of neighboring sound sources, especially in the lateral acoustic space. PMID:23781211

  19. Characterization of space dust using acoustic impact detection.

    PubMed

    Corsaro, Robert D; Giovane, Frank; Liou, Jer-Chyi; Burchell, Mark J; Cole, Michael J; Williams, Earl G; Lagakos, Nicholas; Sadilek, Albert; Anderson, Christopher R

    2016-08-01

    This paper describes studies leading to the development of an acoustic instrument for measuring properties of micrometeoroids and other dust particles in space. The instrument uses a pair of easily penetrated membranes separated by a known distance. Sensors located on these films detect the transient acoustic signals produced by particle impacts. The arrival times of these signals at the sensor locations are used in a simple multilateration calculation to measure the impact coordinates on each film. Particle direction and speed are found using these impact coordinates and the known membrane separations. This ability to determine particle speed, direction, and time of impact provides the information needed to assign the particle's orbit and identify its likely origin. In many cases additional particle properties can be estimated from the signal amplitudes, including approximate diameter and (for small particles) some indication of composition/morphology. Two versions of this instrument were evaluated in this study. Fiber optic displacement sensors are found advantageous when very thin membranes can be maintained in tension (solar sails, lunar surface). Piezoelectric strain sensors are preferred for thicker films without tension (long duration free flyers). The latter was selected for an upcoming installation on the International Space Station. PMID:27586768

  20. Characterization of space dust using acoustic impact detection.

    PubMed

    Corsaro, Robert D; Giovane, Frank; Liou, Jer-Chyi; Burchell, Mark J; Cole, Michael J; Williams, Earl G; Lagakos, Nicholas; Sadilek, Albert; Anderson, Christopher R

    2016-08-01

    This paper describes studies leading to the development of an acoustic instrument for measuring properties of micrometeoroids and other dust particles in space. The instrument uses a pair of easily penetrated membranes separated by a known distance. Sensors located on these films detect the transient acoustic signals produced by particle impacts. The arrival times of these signals at the sensor locations are used in a simple multilateration calculation to measure the impact coordinates on each film. Particle direction and speed are found using these impact coordinates and the known membrane separations. This ability to determine particle speed, direction, and time of impact provides the information needed to assign the particle's orbit and identify its likely origin. In many cases additional particle properties can be estimated from the signal amplitudes, including approximate diameter and (for small particles) some indication of composition/morphology. Two versions of this instrument were evaluated in this study. Fiber optic displacement sensors are found advantageous when very thin membranes can be maintained in tension (solar sails, lunar surface). Piezoelectric strain sensors are preferred for thicker films without tension (long duration free flyers). The latter was selected for an upcoming installation on the International Space Station.

  1. Audiovisual Vowel Monitoring and the Word Superiority Effect in Children

    ERIC Educational Resources Information Center

    Fort, Mathilde; Spinelli, Elsa; Savariaux, Christophe; Kandel, Sonia

    2012-01-01

    The goal of this study was to explore whether viewing the speaker's articulatory gestures contributes to lexical access in children (ages 5-10) and in adults. We conducted a vowel monitoring task with words and pseudo-words in audio-only (AO) and audiovisual (AV) contexts with white noise masking the acoustic signal. The results indicated that…

  2. Effects of Talker Variability on Vowel Recognition in Cochlear Implants

    ERIC Educational Resources Information Center

    Chang, Yi-ping; Fu, Qian-Jie

    2006-01-01

    Purpose: To investigate the effects of talker variability on vowel recognition by cochlear implant (CI) users and by normal-hearing (NH) participants listening to 4-channel acoustic CI simulations. Method: CI users were tested with their clinically assigned speech processors. For NH participants, 3 CI processors were simulated, using different…

  3. Thresholds for second formant transitions in front vowels

    NASA Astrophysics Data System (ADS)

    Kewley-Port, Diane; Goodman, Shawn S.

    2005-11-01

    Formant dynamics in vowel nuclei contribute to vowel classification in English. This study examined listeners' ability to discriminate dynamic second formant transitions in synthetic high front vowels. Acoustic measurements were made from the nuclei (steady state and 20% and 80% of vowel duration) for the vowels /i,I,e,ɛ,æ/ spoken by a female in /bVd/ context. Three synthesis parameters were selected to yield twelve discrimination conditions: initial frequency value for F2 (2525, 2272, or 2068 Hz), slope direction (rising or falling), and duration (110 or 165 ms). F1 frequency was roved. In the standard stimuli, F0 and F1-F4 were steady state. In the comparison stimuli only F2 frequency varied linearly to reach a final frequency. Five listeners were tested under adaptive tracking to estimate the threshold for frequency extent, the minimal detectable difference in frequency between the initial and final F2 values, called ΔF extent. Analysis showed that initial F2 frequency and direction of movement for some F2 frequencies contributed to significant differences in ΔF extent. Results suggested that listeners attended to differences in the stimulus property of frequency extent (hertz), not formant slope (hertz/second). Formant extent thresholds were at least four times smaller than extents measured in the natural speech tokens, and 18 times smaller than for the diphthongized vowel /e/.

  4. Effects of local lexical competition and regional dialect on vowel production.

    PubMed

    Clopper, Cynthia G; Tamati, Terrin N

    2014-07-01

    Global measures of lexical competition, such as lexical neighborhood density, assume that all phonological contrasts contribute equally to competition. However, effects of local phonetic similarity have also been observed in speech production processes, suggesting that some contrasts may lead to greater competition than others. In the current study, the effect of local lexical competition on vowel production was examined across two dialects of American English that differ in the phonetic similarity of the low-front and low-back vowel pairs. Results revealed a significant interaction between regional dialect and local lexical competition on the acoustic distance within each vowel pair. Local lexical contrast led to greater acoustic distance between vowels, as expected, but this effect was significantly enhanced for acoustically similar dialect-specific variants. These results were independent of global neighborhood density, suggesting that local lexical competition may contribute to the realization of sociolinguistic variation and phonological change.

  5. A New Acoustic Test Facility at Alcatel Space Test Centre

    NASA Astrophysics Data System (ADS)

    Meurat, A.; Jezequel, L.

    2004-08-01

    Due to the obsolescence of its acoustic test facility, Alcatel Space has initiated the investment of a large acoustic chamber on its test centre located in Cannes, south of France. This paper presents the main specification elaborated to design the facility, and the solution chosen : it will be located on a dedicated area of the existing test centre and will be based on technical solution already used in similar facilities over the world. The main structure consists in a chamber linked to an external envelope (concrete building) through suspension aiming at decoupling the vibration and preventing from seismic risks. The noise generation system is based on the use of Wyle modulators located on the chamber roof. Gaseous nitrogen is produced by a dedicated gas generator developed by Air-Liquide that could deliver high flow rate with accurate pressure and temperature controls. The control and acquisition system is based on existing solution implemented on the vibration facilities of the test centre. With the start of the construction in May 2004, the final acceptance tests are planned for April 2005, and the first satellites to be tested are planned for May 2005.

  6. Durational and spectral differences in American English vowels: dialect variation within and across regions.

    PubMed

    Fridland, Valerie; Kendall, Tyler; Farrington, Charlie

    2014-07-01

    Spectral differences among varieties of American English have been widely studied, typically recognizing three major regionally diagnostic vowel shift patterns [Labov, Ash, and Boberg (2006). The Atlas of North American English: Phonetics, Phonology and Sound Change (De Gruyter, Berlin)]. Durational variability across dialects, on the other hand, has received relatively little attention. This paper investigates to what extent regional differences in vowel duration are linked with spectral changes taking place in the Northern, Western, and Southern regions of the U.S. Using F1/F2 and duration measures, the durational correlates of the low back vowel merger, characteristic of Western dialects, and the acoustic reversals of the front tense/lax vowels, characteristic of Southern dialects, are investigated. Results point to a positive correlation between spectral overlap and vowel duration for Northern and Western speakers, suggesting that both F1/F2 measures and durational measures are used for disambiguation of vowel quality. The findings also indicate that, regardless of region, a durational distinction maintains the contrast between the low back vowel classes, particularly in cases of spectral merger. Surprisingly, Southerners show a negative correlation for the vowel shifts most defining of contemporary Southern speech, suggesting that neither spectral position nor durational measures are the most relevant cues for vowel quality in the South.

  7. Are vowel errors influenced by consonantal context in the speech of persons with aphasia?

    NASA Astrophysics Data System (ADS)

    Gelfer, Carole E.; Bell-Berti, Fredericka; Boyle, Mary

    2001-05-01

    The literature suggests that vowels and consonants may be affected differently in the speech of persons with conduction aphasia (CA) or nonfluent aphasia with apraxia of speech (AOS). Persons with CA have shown similar error rates across vowels and consonants, while those with AOS have shown more errors for consonants than vowels. These data have been interpreted to suggest that consonants have greater gestural complexity than vowels. However, recent research [M. Boyle et al., Proc. International Cong. Phon. Sci., 3265-3268 (2003)] does not support this interpretation: persons with AOS and CA both had a high proportion of vowel errors, and vowel errors almost always occurred in the context of consonantal errors. To examine the notion that vowels are inherently less complex than consonants and are differentially affected in different types of aphasia, vowel production in different consonantal contexts for speakers with AOS or CA was examined. The target utterances, produced in carrier phrases, were bVC and bV syllables, allowing us to examine whether vowel production is influenced by consonantal context. Listener judgments were obtained for each token, and error productions were grouped according to the intended utterance and error type. Acoustical measurements were made from spectrographic displays.

  8. Final Vowel-Consonant-e.

    ERIC Educational Resources Information Center

    Burmeister, Lou E.

    The utility value of the final vowel-consonant-e phonic generalization was examined using 2,715 common English words. When the vowel was defined as a single-vowel, the consonant as a single-consonant, and the final e as a single-e the generalization was found to be highly useful, contrary to other recent findings. Using the total sample of 2,715…

  9. Vowel reduction across tasks for male speakers of American English.

    PubMed

    Kuo, Christina; Weismer, Gary

    2016-07-01

    This study examined acoustic variation of vowels within speakers across speech tasks. The overarching goal of the study was to understand within-speaker variation as one index of the range of normal speech motor behavior for American English vowels. Ten male speakers of American English performed four speech tasks including citation form sentence reading with a clear-speech style (clear-speech), citation form sentence reading (citation), passage reading (reading), and conversational speech (conversation). Eight monophthong vowels in a variety of consonant contexts were studied. Clear-speech was operationally defined as the reference point for describing variation. Acoustic measures associated with the conventions of vowel targets were obtained and examined. These included temporal midpoint formant frequencies for the first three formants (F1, F2, and F3) and the derived Euclidean distances in the F1-F2 and F2-F3 planes. Results indicated that reduction toward the center of the F1-F2 and F2-F3 planes increased in magnitude across the tasks in the order of clear-speech, citation, reading, and conversation. The cross-task variation was comparable for all speakers despite fine-grained individual differences. The characteristics of systematic within-speaker acoustic variation across tasks have potential implications for the understanding of the mechanisms of speech motor control and motor speech disorders. PMID:27475161

  10. Call Me Alix, Not Elix: Vowels Are More Important than Consonants in Own-Name Recognition at 5 Months

    ERIC Educational Resources Information Center

    Bouchon, Camillia; Floccia, Caroline; Fux, Thibaut; Adda-Decker, Martine; Nazzi, Thierry

    2015-01-01

    Consonants and vowels differ acoustically and articulatorily, but also functionally: Consonants are more relevant for lexical processing, and vowels for prosodic/syntactic processing. These functional biases could be powerful bootstrapping mechanisms for learning language, but their developmental origin remains unclear. The relative importance of…

  11. The Interplay between Input and Initial Biases: Asymmetries in Vowel Perception during the First Year of Life

    ERIC Educational Resources Information Center

    Pons, Ferran; Albareda-Castellot, Barbara; Sebastian-Galles, Nuria

    2012-01-01

    Vowels with extreme articulatory-acoustic properties act as natural referents. Infant perceptual asymmetries point to an underlying bias favoring these referent vowels. However, as language experience is gathered, distributional frequency of speech sounds could modify this initial bias. The perception of the /i/-/e/ contrast was explored in 144…

  12. Space Launch System Scale Model Acoustic Test Ignition Overpressure Testing

    NASA Technical Reports Server (NTRS)

    Nance, Donald; Liever, Peter; Nielsen, Tanner

    2015-01-01

    The overpressure phenomenon is a transient fluid dynamic event occurring during rocket propulsion system ignition. This phenomenon results from fluid compression of the accelerating plume gas, subsequent rarefaction, and subsequent propagation from the exhaust trench and duct holes. The high-amplitude unsteady fluid-dynamic perturbations can adversely affect the vehicle and surrounding structure. Commonly known as ignition overpressure (IOP), this is an important design-to environment for the Space Launch System (SLS) that NASA is currently developing. Subscale testing is useful in validating and verifying the IOP environment. This was one of the objectives of the Scale Model Acoustic Test, conducted at Marshall Space Flight Center. The test data quantifies the effectiveness of the SLS IOP suppression system and improves the analytical models used to predict the SLS IOP environments. The reduction and analysis of the data gathered during the SMAT IOP test series requires identification and characterization of multiple dynamic events and scaling of the event waveforms to provide the most accurate comparisons to determine the effectiveness of the IOP suppression systems. The identification and characterization of the overpressure events, the waveform scaling, the computation of the IOP suppression system knockdown factors, and preliminary comparisons to the analytical models are discussed.

  13. Space Launch System Scale Model Acoustic Test Ignition Overpressure Testing

    NASA Technical Reports Server (NTRS)

    Nance, Donald K.; Liever, Peter A.

    2015-01-01

    The overpressure phenomenon is a transient fluid dynamic event occurring during rocket propulsion system ignition. This phenomenon results from fluid compression of the accelerating plume gas, subsequent rarefaction, and subsequent propagation from the exhaust trench and duct holes. The high-amplitude unsteady fluid-dynamic perturbations can adversely affect the vehicle and surrounding structure. Commonly known as ignition overpressure (IOP), this is an important design-to environment for the Space Launch System (SLS) that NASA is currently developing. Subscale testing is useful in validating and verifying the IOP environment. This was one of the objectives of the Scale Model Acoustic Test (SMAT), conducted at Marshall Space Flight Center (MSFC). The test data quantifies the effectiveness of the SLS IOP suppression system and improves the analytical models used to predict the SLS IOP environments. The reduction and analysis of the data gathered during the SMAT IOP test series requires identification and characterization of multiple dynamic events and scaling of the event waveforms to provide the most accurate comparisons to determine the effectiveness of the IOP suppression systems. The identification and characterization of the overpressure events, the waveform scaling, the computation of the IOP suppression system knockdown factors, and preliminary comparisons to the analytical models are discussed.

  14. Two cross-linguistic factors underlying tongue shapes for vowels

    SciTech Connect

    Nix, D.A.; Papcun, G.; Hogden, J.; Zlokarnik, I.

    1996-06-01

    Desirable characteristics of a vocal-tract parametrization include accuracy, low dimensionality, and generalizability across speakers and languages. A low-dimensional, speaker-independent linear parametrization of vowel tongue shapes can be obtained using the PARAFAC three-mode factor analysis procedure. Harshman et al. applied PARAFAC to midsagittal x-ray vowel data from five English speakers, reporting that two speaker-independent factors are required to accurately represent the tongue shape measured along anatomically normalized vocal-tract diameter grid lines. Subsequently, the cross-linguistic generality of this parametrization was brought into question by the application of PARAFAC to Icelandic vowel data, where three nonorthogonal factors were reported. This solution is shown to be degenerate; a reanalysis of Jackson`s Icelandic data produces two factors that match Harshman et al.`s factors for English vowels, contradicting Jackson`s distinction between English and Icelandic language-specific `articulatory primes.` To obtain vowel factors not constrained by artificial measurement grid lines, x-ray tongue shape traces of six English speakers were marked with 13 equally spaced points. PARAFAC analysis of this unconstranied (x,y) coordinate data results in two factors that are clearly interpretable in terms of the traditional vowel quality dimensions front/back, high/low. 14 refs., 8 figs., 2 tabs.

  15. Training Japanese listeners to identify American English vowels

    NASA Astrophysics Data System (ADS)

    Nishi, Kanae; Kewley-Port, Diane

    2005-04-01

    Perception training of phonemes by second language (L2) learners has been studied primarily using consonant contrasts, where the number of contrasting sounds rarely exceeds five. In order to investigate the effects of stimulus sets, this training study used two conditions: 9 American English vowels covering the entire vowel space (9V), and 3 difficult vowels for problem-focused training (3V). Native speakers of Japanese were trained for nine days. To assess changes in performance due to training, a battery of perception and production tests were given pre- and post-training, as well as 3 months following training. The 9V trainees improved vowel perception on all vowels after training, on average by 23%. Their performance at the 3-month test was slightly worse than the posttest, but still better than the pretest. Transfer of training effect to stimuli spoken by new speakers was observed. Strong response bias observed in the pretest disappeared after the training. The preliminary results of the 3V trainees showed substantial improvement only on the trained vowels. The implications of this research for improved training of L2 learners to understand speech will be discussed. [Work supported by NIH-NIDCD DC-006313 & DC-02229.

  16. Perception of steady-state vowels and vowelless syllables by adults and children

    NASA Astrophysics Data System (ADS)

    Nittrouer, Susan

    2005-04-01

    Vowels can be produced as long, isolated, and steady-state, but that is not how they are found in natural speech. Instead natural speech consists of almost continuously changing (i.e., dynamic) acoustic forms from which mature listeners recover underlying phonetic form. Some theories suggest that children need steady-state information to recognize vowels (and so learn vowel systems), even though that information is sparse in natural speech. The current study examined whether young children can recover vowel targets from dynamic forms, or whether they need steady-state information. Vowel recognition was measured for adults and children (3, 5, and 7 years) for natural productions of /dæd/, /dUd/ /æ/, /U/ edited to make six stimulus sets: three dynamic (whole syllables; syllables with middle 50-percent replaced by cough; syllables with all but the first and last three pitch periods replaced by cough), and three steady-state (natural, isolated vowels; reiterated pitch periods from those vowels; reiterated pitch periods from the syllables). Adults scored nearly perfectly on all but first/last three pitch period stimuli. Children performed nearly perfectly only when the entire syllable was heard, and performed similarly (near 80%) for all other stimuli. Consequently, children need dynamic forms to perceive vowels; steady-state forms are not preferred.

  17. Lexical Representation of Japanese Vowel Devoicing

    ERIC Educational Resources Information Center

    Ogasawara, Naomi

    2013-01-01

    Vowel devoicing happens in Japanese when the high vowel is between voiceless consonants. The aim of this study is to investigate the lexical representation of vowel devoicing. A long-term repetition-priming experiment was conducted. Participants shadowed words containing either a devoiced or a voiced vowel in three priming paradigms, and their…

  18. The Vowel Harmony in the Sinhala Language

    ERIC Educational Resources Information Center

    Petryshyn, Ivan

    2005-01-01

    The Sinhala language is characterized by the melodic shifty stress or its essence, the opposition between long and short vowels, the Ablaut variants of the vowels and the syllabic alphabet which, of course, might impact the vowel harmony and can be a feature of all the leveled Indo-European languages. The vowel harmony is a well-known concept in…

  19. Sparseness of vowel category structure: Evidence from English dialect comparison.

    PubMed

    Scharinger, Mathias; Idsardi, William J

    2014-02-01

    Current models of speech perception tend to emphasize either fine-grained acoustic properties or coarse-grained abstract characteristics of speech sounds. We argue for a particular kind of 'sparse' vowel representations and provide new evidence that these representations account for the successful access of the corresponding categories. In an auditory semantic priming experiment, American English listeners made lexical decisions on targets (e.g. load) preceded by semantically related primes (e.g. pack). Changes of the prime vowel that crossed a vowel-category boundary (e.g. peck) were not treated as a tolerable variation, as assessed by a lack of priming, although the phonetic categories of the two different vowels considerably overlap in American English. Compared to the outcome of the same experiment with New Zealand English listeners, where such prime variations were tolerated, our experiment supports the view that phonological representations are important in guiding the mapping process from the acoustic signal to an abstract mental representation. Our findings are discussed with regard to current models of speech perception and recent findings from brain imaging research.

  20. Perception of English vowels by bilingual Chinese-English and corresponding monolingual listeners.

    PubMed

    Yang, Jing; Fox, Robert A

    2014-06-01

    This study compares the underlying perceptual structure of vowel perception in monolingual Chinese, monolingual English and bilingual Chinese-English listeners. Of particular interest is how listeners' spatial organization of vowels is affected either by their L1 or their experience with L2. Thirteen English vowels, /i, I, e, epsilon, ae, u, omega, o, (see symbol), alpha, (see symbol)I, alphaI, alphaomega/, embedded in /hVd/ syllable produced by an Ohio male speaker were presented in pairs to three groups of listeners. Each listener rated 312 vowel pairs on a nine-point dissimilarity scale. The responses from each group were analyzed using a multidimensional scaling program (ALSCAL). Results demonstrated that all three groups of listeners used high/low and front/back distinctions as the two most important dimensions to perceive English vowels. However, the vowels were distributed in clusters in the perceptual space of Chinese monolinguals, while they were appropriately separated and located in that of bilinguals and English monolinguals. Besides the two common perceptual dimensions, each group of listeners utilized a different third dimension to perceive these English vowels. English monolinguals used high-front offset. Bilinguals used a dimension mainly correlated to the distinction of monophthong/diphthong. Chinese monolinguals separated two high vowels, /i/ and /u/, from the rest of vowels in the third dimension. The difference between English monolinguals and Chinese monolinguals evidenced the effect of listeners' native language on the vowel perception. The difference between Chinese monolinguals and bilingual listeners as well as the approximation of bilingual listeners' perceptual space to that of English monolinguals demonstrated the effect of L2 experience on listeners' perception of L2 vowels.

  1. A comparison of vowel formant frequencies in the babbling of infants exposed to Canadian English and Canadian French

    NASA Astrophysics Data System (ADS)

    Mattock, Karen; Rvachew, Susan; Polka, Linda; Turner, Sara

    2005-04-01

    It is well established that normally developing infants typically enter the canonical babbling stage of production between 6 and 8 months of age. However, whether the linguistic environment affects babbling, either in terms of the phonetic inventory of vowels produced by infants [Oller & Eiler (1982)] or the acoustics of vowel formants [Boysson-Bardies et al. (1989)] is controversial. The spontaneous speech of 42 Canadian English- and Canadian French-learning infants aged 8 to 11, 12 to 15 and 16 to 18 months of age was recorded and digitized to yield a total of 1253 vowels that were spectrally analyzed and statistically compared for differences in first and second formant frequencies. Language-specific influences on vowel acoustics were hypothesized. Preliminary results reveal changes in formant frequencies as a function of age and language background. There is evidence of decreases over age in the F1 values of French but not English infants vowels, and decreases over age in the F2 values of English but not French infants vowels. The notion of an age-related shift in infants attention to language-specific acoustic features and the implications of this for early vocal development as well as for the production of Canadian English and Canadian French vowels will be discussed.

  2. Perception of vowel length by Japanese- and English-learning infants.

    PubMed

    Mugitani, Ryoko; Pons, Ferran; Fais, Laurel; Dietrich, Christiane; Werker, Janet F; Amano, Shigeaki

    2009-01-01

    This study investigated vowel length discrimination in infants from 2 language backgrounds, Japanese and English, in which vowel length is either phonemic or nonphonemic. Experiment 1 revealed that English 18-month-olds discriminate short and long vowels although vowel length is not phonemically contrastive in English. Experiments 2 and 3 revealed that Japanese 18-month-olds also discriminate the pairs but in an asymmetric manner: They detected only the change from long to short vowel, but not the change in the opposite direction, although English infants in Experiment 1 detected the change in both directions. Experiment 4 tested Japanese 10-month-olds and revealed a symmetric pattern of discrimination similar to that of English 18-month-olds. Experiment 5 revealed that native adult Japanese speakers, unlike Japanese 18-month-old infants who are presumably still developing phonological perception, ultimately acquire a symmetrical discrimination pattern for the vowel contrasts. Taken together, our findings suggest that English 18-month-olds and Japanese 10-month-olds perceive vowel length using simple acoustic?phonetic cues, whereas Japanese 18-month-olds perceive it under the influence of the emerging native phonology, which leads to a transient asymmetric pattern in perception.

  3. Acoustic vibration analysis for utilization of woody plant in space environment

    NASA Astrophysics Data System (ADS)

    Chida, Yukari; Yamashita, Masamichi; Hashimoto, Hirofumi; Sato, Seigo; Tomita-Yokotani, Kaori; Baba, Keiichi; Suzuki, Toshisada; Motohashi, Kyohei; Sakurai, Naoki; Nakagawa-izumi, Akiko

    2012-07-01

    We are proposing to raise woody plants for space agriculture in Mars. Space agriculture has the utilization of wood in their ecosystem. Nobody knows the real tree shape grown under space environment under the low or micro gravitational conditions such as outer environment. Angiosperm tree forms tension wood for keeping their shape. Tension wood formation is deeply related to gravity, but the details of the mechanism of its formation has not yet been clarified. For clarifying the mechanism, the space experiment in international space station, ISS is the best way to investigate about them as the first step. It is necessary to establish the easy method for crews who examine the experiments at ISS. Here, we are proposing to investigate the possibility of the acoustic vibration analysis for the experiment at ISS. Two types of Japanese cherry tree, weeping and upright types in Prunus sp., were analyzed by the acoustic vibration method. Coefficient-of-variation (CV) of sound speed was calculated by the acoustic vibration analysis. The amount of lignin and decomposed lignin were estimated by both Klason and Py-GC/MS method, respectively. The relationships of the results of acoustic vibration analysis and the inner components in tested woody materials were investigated. After the experiments, we confirm the correlation about them. Our results indicated that the acoustic vibration analysis would be useful for determining the inside composition as a nondestructive method in outer space environment.

  4. Vowel perception by noise masked normal-hearing young adults

    NASA Astrophysics Data System (ADS)

    Richie, Carolyn; Kewley-Port, Diane; Coughlin, Maureen

    2005-08-01

    This study examined vowel perception by young normal-hearing (YNH) adults, in various listening conditions designed to simulate mild-to-moderate sloping sensorineural hearing loss. YNH listeners were individually age- and gender-matched to young hearing-impaired (YHI) listeners tested in a previous study [Richie et al., J. Acoust. Soc. Am. 114, 2923-2933 (2003)]. YNH listeners were tested in three conditions designed to create equal audibility with the YHI listeners; a low signal level with and without a simulated hearing loss, and a high signal level with a simulated hearing loss. Listeners discriminated changes in synthetic vowel tokens /smcapi e ɛ invv æ/ when F1 or F2 varied in frequency. Comparison of YNH with YHI results failed to reveal significant differences between groups in terms of performance on vowel discrimination, in conditions of similar audibility by using both noise masking to elevate the hearing thresholds of the YNH and applying frequency-specific gain to the YHI listeners. Further, analysis of learning curves suggests that while the YHI listeners completed an average of 46% more test blocks than YNH listeners, the YHI achieved a level of discrimination similar to that of the YNH within the same number of blocks. Apparently, when age and gender are closely matched between young hearing-impaired and normal-hearing adults, performance on vowel tasks may be explained by audibility alone.

  5. Perceptual effects of dialectal and prosodic variation in vowels

    NASA Astrophysics Data System (ADS)

    Fox, Robert Allen; Jacewicz, Ewa; Hatcher, Kristin; Salmons, Joseph

    2005-09-01

    As was reported earlier [Fox et al., J. Acoust. Soc. Am. 114, 2396 (2003)], certain vowels in the Ohio and Wisconsin dialects of American English are shifting in different directions. In addition, we have found that the spectral characteristics of these vowels (e.g., duration and formant frequencies) changed systematically under varying degrees of prosodic prominence, with somewhat different changes occurring within each dialect. The question addressed in the current study is whether naive listeners from these two dialects are sensitive to both the dialect variations and to the prosodically induced spectral differences. Listeners from Ohio and Wisconsin listened to the stimulus tokens [beIt] and [bɛt] produced in each of three prosodic contexts (representing three different levels of prominence). These words were produced by speakers from Ohio or from Wisconsin (none of the listeners were also speakers). Listeners identified the stimulus tokens in terms of vowel quality and indicated whether it was a good, fair, or poor exemplar of that phonetic category. Results showed that both phonetic quality decisions and goodness ratings were systematically and significantly affected by speaker dialect, listener dialect, and prosodic context. Implications of source and nature of ongoing vowel changes in these two dialects will be discussed. [Work partially supported by NIDCD R03 DC005560-01.

  6. Preliminary characterization of a one-axis acoustic system. [acoustic levitation for space processing

    NASA Technical Reports Server (NTRS)

    Oran, W. A.; Reiss, D. A.; Berge, L. H.; Parker, H. W.

    1979-01-01

    The acoustic fields and levitation forces produced along the axis of a single-axis resonance system were measured. The system consisted of a St. Clair generator and a planar reflector. The levitation force was measured for bodies of various sizes and geometries (i.e., spheres, cylinders, and discs). The force was found to be roughly proportional to the volume of the body until the characteristic body radius reaches approximately 2/k (k = wave number). The acoustic pressures along the axis were modeled using Huygens principle and a method of imaging to approximate multiple reflections. The modeled pressures were found to be in reasonable agreement with those measured with a calibrated microphone.

  7. Technical Aspects of Acoustical Engineering for the ISS [International Space Station

    NASA Technical Reports Server (NTRS)

    Allen, Christopher S.

    2009-01-01

    It is important to control acoustic levels on manned space flight vehicles and habitats to protect crew-hearing, allow for voice communications, and to ensure a healthy and habitable environment in which to work and live. For the International Space Station (ISS) this is critical because of the long duration crew-stays of approximately 6-months. NASA and the JSC Acoustics Office set acoustic requirements that must be met for hardware to be certified for flight. Modules must meet the NC-50 requirement and other component hardware are given smaller allocations to meet. In order to meet these requirements many aspects of noise generation and control must be considered. This presentation has been developed to give an insight into the various technical activities performed at JSC to ensure that a suitable acoustic environment is provided for the ISS crew. Examples discussed include fan noise, acoustic flight material development, on-orbit acoustic monitoring, and a specific hardware development and acoustical design case, the ISS Crew Quarters.

  8. Changes in Wisconsin English over 110 Years: A Real-Time Acoustic Account

    ERIC Educational Resources Information Center

    Delahanty, Jennifer

    2011-01-01

    The growing set of studies on American regional dialects have to date focused heavily on vowels while few examine consonant features and none provide acoustic analysis of both vowel and consonant features. This dissertation uses real-time data on both vowels and consonants to show how Wisconsin English has changed over time. Together, the…

  9. Vowel production in Korean, Korean-accented English, and American English

    NASA Astrophysics Data System (ADS)

    Lee, Jimin; Weismer, Gary

    2005-09-01

    The current study compares vowel formant frequencies and durations produced by ten native speakers of Korean, those same speakers producing American English vowels, and ten native speakers of American English. The Korean speakers were chosen carefully to have a minimum of 2 years, and maximum of 5 years residence in the United States; all speakers were between the ages of 22 and 27. In addition, the native speakers of Korean were chosen, by means of a small-scale dialect-severity experiment, from a larger pool of speakers to achieve some homogeneity in their mastery of English phonetics. The full vowel systems of both languages were explored, and a rate condition was included (conversational versus fast) to test the hypothesis that the English vowel space is modified by rate differently for native speakers of Korean who produce English, versus native speakers of English. Results will be discussed in terms of language- and rate-induced adjustments of the vowel systems under study.

  10. International Space Station USOS Crew Quarters Ventilation and Acoustic Design Implementation

    NASA Technical Reports Server (NTRS)

    Broyan, James Lee, Jr.

    2009-01-01

    The International Space Station (ISS) United States Operational Segment (USOS) has four permanent rack sized ISS Crew Quarters (CQ) providing a private crewmember space. The CQ uses Node 2 cabin air for ventilation/thermal cooling, as opposed to conditioned ducted air from the ISS Temperature Humidity Control System or the ISS fluid cooling loop connections. Consequently, CQ can only increase the air flow rate to reduce the temperature delta between the cabin and the CQ interior. However, increasing airflow causes increased acoustic noise so efficient airflow distribution is an important design parameter. The CQ utilized a two fan push-pull configuration to ensure fresh air at the crewmember s head position and reduce acoustic exposure. The CQ interior needs to be below Noise Curve 40 (NC-40). The CQ ventilation ducts are open to the significantly louder Node 2 cabin aisle way which required significantly acoustic mitigation controls. The design implementation of the CQ ventilation system and acoustic mitigation are very inter-related and require consideration of crew comfort balanced with use of interior habitable volume, accommodation of fan failures, and possible crew uses that impact ventilation and acoustic performance. This paper illustrates the types of model analysis, assumptions, vehicle interactions, and trade-offs required for CQ ventilation and acoustics. Additionally, on-orbit ventilation system performance and initial crew feedback is presented. This approach is applicable to any private enclosed space that the crew will occupy.

  11. Speech intelligibility, speaking rate, and vowel formant characteristics in Mandarin-speaking children with cochlear implant.

    PubMed

    Chuang, Hsiu-Feng; Yang, Cheng-Chieh; Chi, Lin-Yang; Weismer, Gary; Wang, Yu-Tsai

    2012-04-01

    The effects of the use of cochlear implant (CI) on speech intelligibility, speaking rate, and vowel formant characteristics and the relationships between speech intelligibility, speaking rate, and vowel formant characteristics for children are clinically important. The purposes of this study were to report on the comparisons for speaking rate and vowel space area, and their relationship with speech intelligibility, between 24 Mandarin-speaking children with CI and 24 age-sex-education level matched normal hearing (NH) controls. Participants were audio recorded as they read a designed Mandarin intelligibility test, repeated prolongation of each of the three point vowels /i/, /a/, and /u/ five times, and repeated each of three sentences carrying one point vowel five times. Compared to the NH group, the CI group exhibited: (1) mild-to-moderate speech intelligibility impairment; (2) significantly reduced speaking rate mainly due to significantly longer inter-word pauses and larger pause proportion; and (3) significantly less vowel reduction in the horizontal dimension in sustained vowel phonation. The limitations of speech intelligibility development in children after cochlear implantation were related to atypical patterns and to a smaller degree in vowel reduction and slower speaking rate resulting from less efficient articulatory movement transition.

  12. Vowel category dependence of the relationship between palate height, tongue height, and oral area.

    PubMed

    Hasegawa-Johnson, Mark; Pizza, Shamala; Alwan, Abeer; Cha, Jul Setsu; Haker, Katherine

    2003-06-01

    This article evaluates intertalker variance of oral area, logarithm of the oral area, tongue height, and formant frequencies as a function of vowel category. The data consist of coronal magnetic resonance imaging (MRI) sequences and acoustic recordings of 5 talkers, each producing 11 different vowels. Tongue height (left, right, and midsagittal), palate height, and oral area were measured in 3 coronal sections anterior to the oropharyngeal bend and were subjected to multivariate analysis of variance, variance ratio analysis, and regression analysis. The primary finding of this article is that oral area (between palate and tongue) showed less intertalker variance during production of vowels with an oral place of articulation (palatal and velar vowels) than during production of vowels with a uvular or pharyngeal place of articulation. Although oral area variance is place dependent, percentage variance (log area variance) is not place dependent. Midsagittal tongue height in the molar region was positively correlated with palate height during production of palatal vowels, but not during production of nonpalatal vowels. Taken together, these results suggest that small oral areas are characterized by relatively talker-independent vowel targets and that meeting these talker-independent targets is important enough that each talker adjusts his or her own tongue height to compensate for talker-dependent differences in constriction anatomy. Computer simulation results are presented to demonstrate that these results may be explained by an acoustic control strategy: When talkers with very different anatomical characteristics try to match talker-independent formant targets, the resulting area variances are minimized near the primary vocal tract constriction. PMID:14697000

  13. Interaction of native- and second-language vowel system(s) in early and late bilinguals.

    PubMed

    Baker, Wendy; Trofimovich, Pavel

    2005-01-01

    The objective of this study was to determine how bilinguals' age at the time of language acquisition influenced the organization of their phonetic system(s). The productions of six English and five Korean vowels by English and Korean monolinguals were compared to the productions of the same vowels by early and late Korean-English bilinguals varying in amount of exposure to their second language. Results indicated that bilinguals' age profoundly influenced both the degree and the direction of the interaction between the phonetic systems of their native (L1) and second (L2) languages. In particular, early bilinguals manifested a bidirectional L1-L2 influence and produced distinct acoustic realizations of L1 and L2 vowels. Late bilinguals, however, showed evidence of a unidirectional influence of the L1 on the L2 and produced L2 vowels that were "colored" by acoustic properties of their L1. The degree and direction of L1-L2 influences in early and late bilinguals appeared to depend on the degree of acoustic similarity between L1 and L2 vowels and the length of their exposure to the L2. Overall, the findings underscored the complex nature of the restructuring of the L1-L2 phonetic system(s) in bilinguals. PMID:16161470

  14. Fan Acoustic Issues in the NASA Space Flight Experience

    NASA Technical Reports Server (NTRS)

    Allen, Christopher S.; Goodman, Jerry

    2008-01-01

    Emphasis needs to be placed on choosing quiet fans compatible with systems design and specifications that control spec levels: a) Sound power; b) Choose quiet fan or plan to quiet it, early in program; c) Plan early verification that fan source allocations are met. Airborne noise: a) System design should function/play together with fans used (flow passages, restrictions, bends, expansions & contractions, and acoustics) vs. fan speed understood (nominal, worst case, & unplanned variances); b) Fan inlets treated, as required; c) Fan Outlets treated, as required; d) Ducted system inlets are outlets designed for acoustic compliance compatibility & designed so some late required modifications can be made without significant impacts. Structure Borne Noise: a) Structure borne noise dealt with as part of fan package or installation; b) Duct attachments and lines isolated. Case Radiated Noise: - Treatment added as much as possible to fan package (see example).

  15. Acoustic response analysis of large light space structures

    NASA Astrophysics Data System (ADS)

    Defosse, H.; Mercier, F.

    1989-10-01

    The dynamic behavior of large lightweight aerospace structures under reverberant acoustic excitation is investigated. A review of the modal superposition theory is presented, along with an improved analysis method of air mass and acoustic radiation damping effects. An efficient postprocessor uses classic finite element codes to compute structural responses up to medium frequencies. Experiments performed on a honeycomb panel demonstrate the importance of two factors for the accurate analysis of the vibroacoustic responses of such aerospace structures: specifically, it is shown that the low frequency response calculations should include correlation characteristics of the excitation pressure field, and the test data processing should include pressure cross spectra calculations. Theoretical and analytical results are compared to assess air effects on a rigid circular plate. Dynamic analysis of large lightweight aerospace structures under a vacuum hypothesis may lead to a significant overestimation of predicted levels.

  16. Longitudinal developmental changes in spectral peaks of vowels produced by Japanese infants.

    PubMed

    Ishizuka, Kentaro; Mugitani, Ryoko; Kato, Hiroko; Amano, Shigeaki

    2007-04-01

    This paper describes a longitudinal analysis of the vowel development of two Japanese infants in terms of spectral resonant peaks. This study aims to investigate when and how the two infants become able to produce categorically separated vowels, and covers the ages of 4 to 60 months in order to provide detailed findings on the developmental process of speech production. The two lower spectral peaks were estimated from vowels extracted from natural spontaneous speech produced by the infants. Phoneme labeled and transcription-independent unlabeled data analyses were conducted. The labeled data analysis revealed longitudinal trends in the developmental change, which correspond to the articulation positions of the tongue and the rapid enlargement of the articulatory organs. In addition, the distribution of the two spectral peaks demonstrates the vowel space expansion that occurs with age. An unlabeled data analysis technique derived from the linear discriminant analysis method was introduced to measure the vowel space expansion quantitatively. It revealed that the infant's vowel space becomes similar to that of an adult in the early stages. In terms of both labeled and unlabeled properties, these results suggested that infants become capable of producing categorically separated vowels by 24 months.

  17. Mixing fuel particles for space combustion research using acoustics

    NASA Technical Reports Server (NTRS)

    Burns, Robert J.; Johnson, Jerome A.; Klimek, Robert B.

    1988-01-01

    Part of the microgravity science to be conducted aboard the Shuttle (STS) involves combustion using solids, particles, and liquid droplets. The central experimental facts needed for characterization of premixed quiescent particle cloud flames cannot be adequately established by normal gravity studies alone. The experimental results to date of acoustically mixing a prototypical particulate, lycopodium, in a 5 cm diameter by 75 cm long flame tube aboard a Learjet aircraft flying a 20 sec low gravity trajectory are described. Photographic and light detector instrumentation combine to measure and characterize particle cloud uniformity.

  18. Auditory, Visual, and Auditory-Visual Perception of Vowels by Hearing-Impaired Children.

    ERIC Educational Resources Information Center

    Hack, Zarita Caplan; Erber, Norman P.

    1982-01-01

    Vowels were presented through auditory, visual, and auditory-visual modalities to 18 hearing impaired children (12 to 15 years old) having good, intermediate, and poor auditory word recognition skills. All the groups had difficulty with acoustic information and visual information alone. The first two groups had only moderate difficulty identifying…

  19. Phonetic correlates of phonological vowel quantity in Yakut read and spontaneous speech.

    PubMed

    Vasilyeva, Lena; Arnhold, Anja; Järvikivi, Juhani

    2016-05-01

    The quantity language Yakut (Sakha) has a binary distinction between short and long vowels. Disyllabic words with short and long vowels in one or both syllables were extracted from spontaneous speech of native Yakut speakers. In addition, a controlled production by a native speaker of disyllabic words with different short and long vowel combinations along with contrastive minimal pairs was recorded in a phonetics laboratory. Acoustic measurements of the vowels' fundamental frequency, duration, and intensity showed a significant consistent lengthening of phonologically long vowels compared to their short counterparts. However, in addition to evident durational differences between long and short quantities, fundamental frequency and intensity also showed effects of quantity. These results allow the interpretation that similarly to other non-tonal quantity languages like Finnish or Estonian, the Yakut vowel quantity opposition is not based exclusively on durational differences. The data furthermore revealed differences in F0 contours between spontaneous and read speech, providing some first indications of utterance-level prosody in Yakut. PMID:27250149

  20. Age-related Changes in Acoustic Modifications of Mandarin Maternal Speech to Preverbal Infants and Five-Year-Old Children: A Longitudinal Study

    PubMed Central

    Liu, Huei-Mei; Tsao, Feng-Ming; Kuhl, Patricia K.

    2010-01-01

    Acoustic-phonetic exaggeration of infant-directed speech (IDS) is well documented, but few studies address whether these features are modified with a child's age. Mandarin-speaking mothers were recorded while addressing an adult and their child at two ages (7-12 months and 5 years) to examine the acoustic-phonetic differences between IDS and child–directed speech (CDS). CDS exhibits an exaggeration pattern resembling that of IDS—expanded vowel space, longer vowels, higher pitch, and greater lexical tone differences—when compared to ADS. Longitudinal analysis demonstrated that the extent of acoustic exaggeration is significantly smaller in CDS than in IDS. Age-related changes in maternal speech provide some support for the hypothesis that mothers adjust their speech directed toward children as a function of the child's language ability. PMID:19232142

  1. Anticipatory consonant-to-vowel coarticulation in the production of VCV sequences in Italian.

    PubMed

    Farnetani, E; Recasens, D

    1993-01-01

    The study of the spatio-temporal interactions between contiguous segments in speech is a means for a better understanding of how segments are serially organized. This research explores the relation between a vowel and the following consonant by studying the anticipatory C-to-V coarticulatory effects by means of electropalatography. The materials were 'VCV utterances produced in isolation and in connected speech by three Italian speakers, with /a/ and /i/ as vowels and the coronals /t, d, l, z, integral of/ as intervocalic consonants. The results show that the consonants affect both the vocal tract configuration of the preceding vowel and its acoustic duration. The spatial effects increase from laterals to stops to fricatives. The tongue body position is raised during /a/ and lowered during /i/. The effects are much larger for /a/ than for /i/ and larger in connected speech than in isolated words. As for temporal coarticulatory effects, the data indicate that vowels tend to be shorter before /integral of/ than before /z/, and shorter before /t/ than before /d/ than before /l/. Spatial and temporal measurements of change in tongue body contact from vowel to the consonantal closure/constriction suggest that the consonants differ among each other in the dynamics as well as in the timing of their gestures, with ampler/longer movements (e.g., for /integral of/) starting earlier than smaller/shorter movements (e.g., for /d/ or /l/). These patterns result in smaller differences between the durations of the total VC sequences than between the individual durations of V or C segments, and suggest that intersegmental organization between vowels and following consonants may have the rhythmic function of reducing the variability of vowel-to-vowel temporal intervals.

  2. Effects of bite blocks and hearing status on vowel production

    NASA Astrophysics Data System (ADS)

    Lane, Harlan; Denny, Margaret; Guenther, Frank H.; Matthies, Melanie L.; Menard, Lucie; Perkell, Joseph S.; Stockmann, Ellen; Tiede, Mark; Vick, Jennell; Zandipour, Majid

    2005-09-01

    This study explores the effects of hearing status and bite blocks on vowel production. Normal-hearing controls and postlingually deaf adults read elicitation lists of /hVd/ syllables with and without bite blocks and auditory feedback. Deaf participants' auditory feedback was provided by a cochlear prosthesis and interrupted by switching off their implant microphones. Recording sessions were held before prosthesis was provided and one month and one year after. Long-term absence of auditory feedback was associated with heightened dispersion of vowel tokens, which was inflated further by inserting bite blocks. The restoration of some hearing with prosthesis reduced dispersion. Deaf speakers' vowel spaces were reduced in size compared to controls. Insertion of bite blocks reduced them further because of the speakers' incomplete compensation. A year of prosthesis use increased vowel contrast with feedback during elicitation. These findings support the inference that models of speech production must assign a role to auditory feedback in error-based correction of feedforward commands for subsequent articulatory gestures.

  3. NONLINEAR BEHAVIOR OF BARYON ACOUSTIC OSCILLATIONS IN REDSHIFT SPACE FROM THE ZEL'DOVICH APPROXIMATION

    SciTech Connect

    McCullagh, Nuala; Szalay, Alexander S.

    2015-01-10

    Baryon acoustic oscillations (BAO) are a powerful probe of the expansion history of the universe, which can tell us about the nature of dark energy. In order to accurately characterize the dark energy equation of state using BAO, we must understand the effects of both nonlinearities and redshift space distortions on the location and shape of the acoustic peak. In a previous paper, we introduced a novel approach to second order perturbation theory in configuration space using the Zel'dovich approximation, and presented a simple result for the first nonlinear term of the correlation function. In this paper, we extend this approach to redshift space. We show how to perform the computation and present the analytic result for the first nonlinear term in the correlation function. Finally, we validate our result through comparison with numerical simulations.

  4. Acoustic levitation technique for containerless processing at high temperatures in space

    NASA Technical Reports Server (NTRS)

    Rey, Charles A.; Merkley, Dennis R.; Hammarlund, Gregory R.; Danley, Thomas J.

    1988-01-01

    High temperature processing of a small specimen without a container has been demonstrated in a set of experiments using an acoustic levitation furnace in the microgravity of space. This processing technique includes the positioning, heating, melting, cooling, and solidification of a material supported without physical contact with container or other surface. The specimen is supported in a potential energy well, created by an acoustic field, which is sufficiently strong to position the specimen in the microgravity environment of space. This containerless processing apparatus has been successfully tested on the Space Shuttle during the STS-61A mission. In that experiment, three samples wer successfully levitated and processed at temperatures from 600 to 1500 C. Experiment data and results are presented.

  5. Predicting midsagittal pharynx shape from tongue position during vowel production.

    PubMed

    Whalen, D H; Kang, A M; Magen, H S; Fulbright, R K; Gore, J C

    1999-06-01

    The shape of the pharynx has a large effect on the acoustics of vowels, but direct measurement of this part of the vocal tract is difficult. The present study examines the efficacy of inferring midsagittal pharynx shape from the position of the tongue, which is much more amenable to measurement. Midsagittal magnetic resonance (MR) images were obtained for multiple repetitions of 11 static English vowels spoken by two subjects (one male and one female). From these, midsagittal widths were measured at approximately 3-mm intervals along the entire vocal tract. A regression analysis was then used to assess whether the pharyngeal widths could be predicted from the locations and width measurements for four positions on the tongue, namely, those likely to be the locations of a receiver coil for an electromagnetometer system. Predictability was quite high throughout the vocal tract (multiple r> 0.9), except for the extreme ends (i.e., larynx and lips) and small decreases for the male subject in the uvula region. The residuals from this analysis showed that the accuracy of predictions was generally quite high, with 89.2% of errors being less than 2 mm. The extremes of the vocal tract, where the resolution of the MRI was poorer, accounted for much of the error. For languages like English, which do not use advanced tongue root (ATR) distinctively, the midsagittal pharynx shape of static vowels can be predicted with high accuracy.

  6. Seeing a voice: Rudolph Koenig's instruments for studying vowel sounds.

    PubMed

    Pantalony, David

    2004-01-01

    The human voice was one of the more elusive acoustical phenomena to study in the 19th century and therefore a crucial test of Hermann von Helmholtz's new theory of sound. This article describes the origins of instruments used to study vowel sounds: synthesizers for production, resonators for detection, and manometric flames for visual display. Instrument maker Rudolph Koenig played a leading role in transforming Helmholtz's ideas into apparatus. In particular, he was the first to make the human voice visible for research and teaching. Koenig's work reveals the rich context of science, craft traditions, experiment, demonstration culture, and commerce in his Paris workshop. PMID:15457810

  7. Measurement and Characterization of Space Shuttle Solid Rocket Motor Plume Acoustics

    NASA Technical Reports Server (NTRS)

    Kenny, Jeremy; Hobbs, Chris; Plotkin, Ken; Pilkey, Debbie

    2009-01-01

    Lift-off acoustic environments generated by the future Ares I launch vehicle are assessed by the NASA Marshall Space Flight Center (MSFC) acoustics team using several prediction tools. This acoustic environment is directly caused by the Ares I First Stage booster, powered by the five-segment Reusable Solid Rocket Motor (RSRMV). The RSRMV is a larger-thrust derivative design from the currently used Space Shuttle solid rocket motor, the Reusable Solid Rocket Motor (RSRM). Lift-off acoustics is an integral part of the composite launch vibration environment affecting the Ares launch vehicle and must be assessed to help generate hardware qualification levels and ensure structural integrity of the vehicle during launch and lift-off. Available prediction tools that use free field noise source spectrums as a starting point for generation of lift-off acoustic environments are described in the monograph NASA SP-8072: "Acoustic Loads Generated by the Propulsion System." This monograph uses a reference database for free field noise source spectrums which consist of subscale rocket motor firings, oriented in horizontal static configurations. The phrase "subscale" is appropriate, since the thrust levels of rockets in the reference database are orders of magnitude lower than the current design thrust for the Ares launch family. Thus, extrapolation is needed to extend the various reference curves to match Ares-scale acoustic levels. This extrapolation process yields a subsequent amount of uncertainty added upon the acoustic environment predictions. As the Ares launch vehicle design schedule progresses, it is important to take every opportunity to lower prediction uncertainty and subsequently increase prediction accuracy. Never before in NASA s history has plume acoustics been measured for large scale solid rocket motors. Approximately twice a year, the RSRM prime vendor, ATK Launch Systems, static fires an assembled RSRM motor in a horizontal configuration at their test facility

  8. Vowel aperture and syllable segmentation in French.

    PubMed

    Goslin, Jeremy; Frauenfelder, Ulrich H

    2008-01-01

    The theories of Pulgram (1970) suggest that if the vowel of a French syllable is open then it will induce syllable segmentation responses that result in the syllable being closed, and vice versa. After the empirical verification that our target French-speaking population was capable of distinguishing between mid-vowel aperture, we examined the relationship between vowel and syllable aperture in two segmentation experiments. Initial findings from a metalinguistic repetition task supported the hypothesis, revealing significant segmentation differences due to vowel aperture across a range of bi-syllabic stimuli. These findings were also supported in an additional online experiment, in which a fragment detection task revealed a syllabic cross-over interaction due to vowel aperture. Evidence from these experiments suggest that multiple, independent cues are used in French syllable segmentation, including vowel aperture.

  9. Prosodic strengthening and featural enhancement: Evidence from acoustic and articulatory realizations of /opena,eye/ in English

    NASA Astrophysics Data System (ADS)

    Cho, Taehong

    2005-06-01

    In this study the effects of accent and prosodic boundaries on the production of English vowels (/opena,eye/), by concurrently examining acoustic vowel formants and articulatory maxima of the tongue, jaw, and lips obtained with EMA (Electromagnetic Articulography) are investigated. The results demonstrate that prosodic strengthening (due to accent and/or prosodic boundaries) has differential effects depending on the source of prominence (in accented syllables versus at edges of prosodic domains; domain initially versus domain finally). The results are interpreted in terms of how the prosodic strengthening is related to phonetic realization of vowel features. For example, when accented, /eye/ was fronter in both acoustic and articulatory vowel spaces (enhancing [-back]), accompanied by an increase in both lip and jaw openings (enhancing sonority). By contrast, at edges of prosodic domains (especially domain-finally), /eye/ was not necessarily fronter, but higher (enhancing [+high]), accompanied by an increase only in the lip (not jaw) opening. This suggests that the two aspects of prosodic structure (accent versus boundary) are differentiated by distinct phonetic patterns. Further, it implies that prosodic strengthening, though manifested in fine-grained phonetic details, is not simply a low-level phonetic event but a complex linguistic phenomenon, closely linked to the enhancement of phonological features and positional strength that may license phonological contrasts. .

  10. From prosodic structure to acoustic saliency: A fMRI investigation of speech rate, clarity, and emphasis

    NASA Astrophysics Data System (ADS)

    Golfinopoulos, Elisa

    Acoustic variability in fluent speech can arise at many stages in speech production planning and execution. For example, at the phonological encoding stage, the grouping of phonemes into syllables determines which segments are coarticulated and, by consequence, segment-level acoustic variation. Likewise phonetic encoding, which determines the spatiotemporal extent of articulatory gestures, will affect the acoustic detail of segments. Functional magnetic resonance imaging (fMRI) was used to measure brain activity of fluent adult speakers in four speaking conditions: fast, normal, clear, and emphatic (or stressed) speech. These speech manner changes typically result in acoustic variations that do not change the lexical or semantic identity of productions but do affect the acoustic saliency of phonemes, syllables and/or words. Acoustic responses recorded inside the scanner were assessed quantitatively using eight acoustic measures and sentence duration was used as a covariate of non-interest in the neuroimaging analysis. Compared to normal speech, emphatic speech was characterized acoustically by a greater difference between stressed and unstressed vowels in intensity, duration, and fundamental frequency, and neurally by increased activity in right middle premotor cortex and supplementary motor area, and bilateral primary sensorimotor cortex. These findings are consistent with right-lateralized motor planning of prosodic variation in emphatic speech. Clear speech involved an increase in average vowel and sentence durations and average vowel spacing, along with increased activity in left middle premotor cortex and bilateral primary sensorimotor cortex. These findings are consistent with an increased reliance on feedforward control, resulting in hyper-articulation, under clear as compared to normal speech. Fast speech was characterized acoustically by reduced sentence duration and average vowel spacing, and neurally by increased activity in left anterior frontal

  11. Effect of Vowel Identity and Onset Asynchrony on Concurrent Vowel Identification

    ERIC Educational Resources Information Center

    Hedrick, Mark S.; Madix, Steven G.

    2009-01-01

    Purpose: The purpose of the current study was to determine the effects of vowel identity and temporal onset asynchrony on identification of vowels overlapped in time. Method: Fourteen listeners with normal hearing, with a mean age of 24 years, participated. The listeners were asked to identify both of a pair of 200-ms vowels (referred to as…

  12. Recreating the real, realizing the imaginary--a composer's preoccupation with acoustic space

    NASA Astrophysics Data System (ADS)

    Godman, Rob

    2002-11-01

    For centuries composers have been concerned with spatialization of sound and with the use of acoustic spaces to create feeling, atmosphere, and musical structure. This paper will explore Rob Godman's own use of sound in space, including (1) his treatment of ancient Vitruvian principles and how they are combined with new technologies; (2) an exploration of virtual journeys through real and imaginary acoustic spaces; (3) how sounds might be perceived in air, liquid, and solids; and (4) how technology has allowed composers to realize ideas that previously had only existed in the imagination. While focusing on artistic concerns, the paper will provide information on research carried out by the composer into acoustic spaces that are able to transform in real time with the aid of digital technology (Max/MSP software with sensor technology) and how these have been used in installation and pre-recorded work. It will also explore digital reconstructions of Vitruvian theatres and how we perceive resonance and ambience in the real and virtual world.

  13. Acoustical Testing Laboratory Developed to Support the Low-Noise Design of Microgravity Space Flight Hardware

    NASA Technical Reports Server (NTRS)

    Cooper, Beth A.

    2001-01-01

    The NASA John H. Glenn Research Center at Lewis Field has designed and constructed an Acoustical Testing Laboratory to support the low-noise design of microgravity space flight hardware. This new laboratory will provide acoustic emissions testing and noise control services for a variety of customers, particularly for microgravity space flight hardware that must meet International Space Station limits on noise emissions. These limits have been imposed by the space station to support hearing conservation, speech communication, and safety goals as well as to prevent noise-induced vibrations that could impact microgravity research data. The Acoustical Testing Laboratory consists of a 23 by 27 by 20 ft (height) convertible hemi/anechoic chamber and separate sound-attenuating test support enclosure. Absorptive 34-in. fiberglass wedges in the test chamber provide an anechoic environment down to 100 Hz. A spring-isolated floor system affords vibration isolation above 3 Hz. These criteria, along with very low design background levels, will enable the acquisition of accurate and repeatable acoustical measurements on test articles, up to a full space station rack in size, that produce very little noise. Removable floor wedges will allow the test chamber to operate in either a hemi/anechoic or anechoic configuration, depending on the size of the test article and the specific test being conducted. The test support enclosure functions as a control room during normal operations but, alternatively, may be used as a noise-control enclosure for test articles that require the operation of noise-generating test support equipment.

  14. The role of vowel quality in stress clash

    NASA Astrophysics Data System (ADS)

    Levey, Sandra Kay

    The effect of stress clash between adjacent primary stressed syllables was examined through perceptual and acoustical analysis. Bisyllabic target words with primary stress on the second syllable were produced in citation form and in sentences by ten adult participants. The selected target words were analyzed for (a)the position of primary stress and (b)the identity of the vowel in the first syllable when produced in citation form. The goal was to determine if primary stress was placed on the final syllable and that the first syllable contained a vowel that could receive primary stress. Words judged to not meet these criteria were eliminated from the corpus. The target words were placed in stress clash contexts (followed by a primary stressed syllable) and in non-clash contexts (followed by a non-primary stressed syllable). The goal was to determine if stress clash resolution would occur, and if so, which of three explanations could account for resolution: (a)stress shift, with primary stress shifted to the first syllable in the target words or stress deletion, with acoustic features reduced in the second syllable in the target words, (b)pitch accent, taking the form of fundamental frequency, assigned to the first syllable in target words produced in early-sentence position or (c)increased final syllable duration in the target word. Perceptual judgment showed that stress clash was resolved inconsistently in stress clash contexts, and that stress shift also occurred in non-clash contexts. Acoustic analysis showed that fundamental frequency was higher the first syllable of target words when stress shift occurred, and that both syllables of the target words were produced with higher fundamental frequency in early-sentence position. A test of the correlation between perceptual judgments and acoustic results showed that fundamental frequency was potentially the primary acoustic feature that signaled the presence of stress shift.

  15. Acoustic correlates of timbre space dimensions: a confirmatory study using synthetic tones.

    PubMed

    Caclin, Anne; McAdams, Stephen; Smith, Bennett K; Winsberg, Suzanne

    2005-07-01

    Timbre spaces represent the organization of perceptual distances, as measured with dissimilarity ratings, among tones equated for pitch, loudness, and perceived duration. A number of potential acoustic correlates of timbre-space dimensions have been proposed in the psychoacoustic literature, including attack time, spectral centroid, spectral flux, and spectrum fine structure. The experiments reported here were designed as direct tests of the perceptual relevance of these acoustical parameters for timbre dissimilarity judgments. Listeners presented with carefully controlled synthetic tones use attack time, spectral centroid, and spectrum fine structure in dissimilarity rating experiments. These parameters thus appear as major determinants of timbre. However, spectral flux appears as a less salient timbre parameter, its salience depending on the number of other dimensions varying concurrently in the stimulus set. Dissimilarity ratings were analyzed with two different multidimensional scaling models (CLASCAL and CONSCAL), the latter providing psychophysical functions constrained by the physical parameters. Their complementarity is discussed. PMID:16119366

  16. Acoustic correlates of timbre space dimensions: a confirmatory study using synthetic tones.

    PubMed

    Caclin, Anne; McAdams, Stephen; Smith, Bennett K; Winsberg, Suzanne

    2005-07-01

    Timbre spaces represent the organization of perceptual distances, as measured with dissimilarity ratings, among tones equated for pitch, loudness, and perceived duration. A number of potential acoustic correlates of timbre-space dimensions have been proposed in the psychoacoustic literature, including attack time, spectral centroid, spectral flux, and spectrum fine structure. The experiments reported here were designed as direct tests of the perceptual relevance of these acoustical parameters for timbre dissimilarity judgments. Listeners presented with carefully controlled synthetic tones use attack time, spectral centroid, and spectrum fine structure in dissimilarity rating experiments. These parameters thus appear as major determinants of timbre. However, spectral flux appears as a less salient timbre parameter, its salience depending on the number of other dimensions varying concurrently in the stimulus set. Dissimilarity ratings were analyzed with two different multidimensional scaling models (CLASCAL and CONSCAL), the latter providing psychophysical functions constrained by the physical parameters. Their complementarity is discussed.

  17. Acoustic Correlates of Emphatic Stress in Central Catalan

    ERIC Educational Resources Information Center

    Nadeu, Marianna; Hualde, Jose Ignacio

    2012-01-01

    A common feature of public speech in Catalan is the placement of prominence on lexically unstressed syllables ("emphatic stress"). This paper presents an acoustic study of radio speech data. Instances of emphatic stress were perceptually identified. Within-word comparison between vowels with emphatic stress and vowels with primary lexical stress…

  18. Haptic Holography: Acoustic Space and the Evolution of the Whole Message

    NASA Astrophysics Data System (ADS)

    Logan, N.

    2013-02-01

    The paper argues that the Haptic Holography Work Station is an example of a medium that fit's with McLuhan's notion of Acoustic Space, that is it is a medium which stimulates more than one sense of perception at a time. As a result, the Haptic Holography Work Station transmits information about the subject much more rapidly than other media that precedes it, be it text, photography or television.

  19. Measurement of the space-time correlation function of thermal acoustic radiation

    NASA Astrophysics Data System (ADS)

    Passechnik, V. I.; Anosov, A. A.; Barabanenkov, Yu. N.; Sel'Sky, A. G.

    2003-09-01

    The space-time correlation function of thermal acoustic radiation pressure is measured for a stationary heated source (a narrow plasticine plate). The correlation dependence is obtained by the multiplication of two signals shifted in time with respect to each other and measured by two receivers. The dependence exhibits an oscillating behavior and changes sign when the source is displaced by half the spatial period of the correlation function.

  20. A wideband fast multipole boundary element method for half-space/plane-symmetric acoustic wave problems

    NASA Astrophysics Data System (ADS)

    Zheng, Chang-Jun; Chen, Hai-Bo; Chen, Lei-Lei

    2013-04-01

    This paper presents a novel wideband fast multipole boundary element approach to 3D half-space/plane-symmetric acoustic wave problems. The half-space fundamental solution is employed in the boundary integral equations so that the tree structure required in the fast multipole algorithm is constructed for the boundary elements in the real domain only. Moreover, a set of symmetric relations between the multipole expansion coefficients of the real and image domains are derived, and the half-space fundamental solution is modified for the purpose of applying such relations to avoid calculating, translating and saving the multipole/local expansion coefficients of the image domain. The wideband adaptive multilevel fast multipole algorithm associated with the iterative solver GMRES is employed so that the present method is accurate and efficient for both lowand high-frequency acoustic wave problems. As for exterior acoustic problems, the Burton-Miller method is adopted to tackle the fictitious eigenfrequency problem involved in the conventional boundary integral equation method. Details on the implementation of the present method are described, and numerical examples are given to demonstrate its accuracy and efficiency.

  1. Perception of cross-language vowel differences: A longitudinal study of native Spanish learners of English

    NASA Astrophysics Data System (ADS)

    Imai, Satomi; Flege, James; Wayland, Ratree

    2002-05-01

    The study evaluated the ability of native Spanish speakers to perceive phonetic differences between Spanish vowels (/i e a o u/) and English vowels (/ieIopen a ocapωu/). Eighteen adult native speakers of Spanish who were learning English as a second language (L2) in Birmingham. AL were tested at 6-month intervals over a 3.5-year period (T1-T7). Five tokens of each Spanish and English vowel were randomly presented for classification in terms of one of the five vowels of Spanish, and were rated for goodness of fit on a 6-point scale (where 0 indicated ``not Spanish'' and 5 indicated a ``good example'' of a Spanish vowel). At both T1 and T7, English /i/, /eI/, /open a /, /ocapω/ and /u/ were usually classified as Spanish /i/, /e/, /a/, /o/ and /u/, respectively. A group analysis revealed that significantly higher ratings were given to the Spanish than English member of the /o/-/ocapω/, /u/-/u/, /e/-/eI/ and /a/-/open a/ pairs but not the /i/-/i/ pair. However, the number of individual L2 learners who gave significantly higher ratings to a Spanish vowel than to the corresponding English vowel differed substantially (/o/-/ocapω/ n=18, /u/-/u/ n=8, /e/-/eI/ n=5, /a/-/open a/ n=1, /i/-/i/n=1). Possible explanations for these between-pair differences -acoustic and perceptual-will be discussed. [Work supported by NIH.

  2. Vowel Quantity and Syllabification in English.

    ERIC Educational Resources Information Center

    Hammond, Michael

    1997-01-01

    Argues that there is phonological gemination in English based on distribution of vowel qualities in medial and final syllables. The analysis, cast in terms of optimality theory, has implications in several domains: (1) ambisyllabicity is not the right way to capture aspiration and flapping; (2) languages in which stress depends on vowel quality…

  3. Preference patterns in infant vowel perception

    NASA Astrophysics Data System (ADS)

    Molnar, Monika T.; Polka, Linda

    2001-05-01

    Infants show directional asymmetries in vowel discrimination tasks that reveal an underlying perceptual bias favoring more peripheral vowels. Polka and Bohn (2003) propose that this bias is language independent and plays an important role in the development of vowel perception. In the present study we measured infant listening preferences for vowels to assess whether a perceptual bias favoring peripheral vowels can be measured more directly. Monolingual (French and English) and bilingual infants completed a listening preference task using multiple natural tokens of German /dut/ and /dyt/ produced by a male talker. In previous work, discrimination of this vowel pair by German-learning and by English-learning infants revealed a robust directional asymmetry in which /u/ acts as a perceptual anchor; specifically, infants had difficulty detecting a change from /u/ to /y/, whereas a change from /y/ to /u/ was readily detected. Preliminary results from preference tests with these stimuli show that most infants between 3 and 5 months of age also listen longer to /u/ than to /y/. Preference data obtained from older infants and with other vowel pairs will also be reported to further test the claim that peripheral vowels have a privileged perceptual status in infant perception.

  4. Vowel identification by cochlear implant users: Contributions of duration cues and dynamic spectral cues

    PubMed Central

    Donaldson, Gail S.; Rogers, Catherine L.; Johnson, Lindsay B.; Oh, Soo Hee

    2015-01-01

    A recent study from our laboratory assessed vowel identification in cochlear implant (CI) users, using full /dVd/ syllables and partial (center- and edges-only) syllables with duration cues neutralized [Donaldson, Rogers, Cardenas, Russell, and Hanna (2013). J. Acoust. Soc. Am. 134, 3021–3028]. CI users' poorer performance for partial syllables as compared to full syllables, and for edges-only syllables as compared to center-only syllables, led to the hypotheses (1) that CI users may rely strongly on vowel duration cues; and (2) that CI users have more limited access to dynamic spectral cues than steady-state spectral cues. The present study tested those hypotheses. Ten CI users and ten young normal hearing (YNH) listeners heard full /dVd/ syllables and modified (center- and edges-only) syllables in which vowel duration cues were either preserved or eliminated. The presence of duration cues significantly improved vowel identification scores in four CI users, suggesting a strong reliance on duration cues. Duration effects were absent for the other CI users and the YNH listeners. On average, CI users and YNH listeners demonstrated similar performance for center-only stimuli and edges-only stimuli having the same total duration of vowel information. However, three CI users demonstrated significantly poorer performance for the edges-only stimuli, indicating apparent deficits of dynamic spectral processing. PMID:26233007

  5. Feedforward and Feedback Control in Apraxia of Speech: Effects of Noise Masking on Vowel Production

    PubMed Central

    Mailend, Marja-Liisa; Guenther, Frank H.

    2015-01-01

    Purpose This study was designed to test two hypotheses about apraxia of speech (AOS) derived from the Directions Into Velocities of Articulators (DIVA) model (Guenther et al., 2006): the feedforward system deficit hypothesis and the feedback system deficit hypothesis. Method The authors used noise masking to minimize auditory feedback during speech. Six speakers with AOS and aphasia, 4 with aphasia without AOS, and 2 groups of speakers without impairment (younger and older adults) participated. Acoustic measures of vowel contrast, variability, and duration were analyzed. Results Younger, but not older, speakers without impairment showed significantly reduced vowel contrast with noise masking. Relative to older controls, the AOS group showed longer vowel durations overall (regardless of masking condition) and a greater reduction in vowel contrast under masking conditions. There were no significant differences in variability. Three of the 6 speakers with AOS demonstrated the group pattern. Speakers with aphasia without AOS did not differ from controls in contrast, duration, or variability. Conclusion The greater reduction in vowel contrast with masking noise for the AOS group is consistent with the feedforward system deficit hypothesis but not with the feedback system deficit hypothesis; however, effects were small and not present in all individual speakers with AOS. Theoretical implications and alternative interpretations of these findings are discussed. PMID:25565143

  6. Identification and Multiplicity of Double Vowels in Cochlear Implant Users

    ERIC Educational Resources Information Center

    Kwon, Bomjun J.; Perry, Trevor T.

    2014-01-01

    Purpose: The present study examined cochlear implant (CI) users' perception of vowels presented concurrently (i.e., "double vowels") to further our understanding of auditory grouping in electric hearing. Method: Identification of double vowels and single vowels was measured with 10 CI subjects. Fundamental frequencies (F0s) of…

  7. Gust Acoustics Computation with a Space-Time CE/SE Parallel 3D Solver

    NASA Technical Reports Server (NTRS)

    Wang, X. Y.; Himansu, A.; Chang, S. C.; Jorgenson, P. C. E.; Reddy, D. R. (Technical Monitor)

    2002-01-01

    The benchmark Problem 2 in Category 3 of the Third Computational Aero-Acoustics (CAA) Workshop is solved using the space-time conservation element and solution element (CE/SE) method. This problem concerns the unsteady response of an isolated finite-span swept flat-plate airfoil bounded by two parallel walls to an incident gust. The acoustic field generated by the interaction of the gust with the flat-plate airfoil is computed by solving the 3D (three-dimensional) Euler equations in the time domain using a parallel version of a 3D CE/SE solver. The effect of the gust orientation on the far-field directivity is studied. Numerical solutions are presented and compared with analytical solutions, showing a reasonable agreement.

  8. Continuous and Discrete Space Particle Filters for Predictions in Acoustic Positioning

    NASA Astrophysics Data System (ADS)

    Bauer, Will; Kim, Surrey; Kouritzin, Michael A.

    2002-12-01

    Predicting the future state of a random dynamic signal based on corrupted, distorted, and partial observations is vital for proper real-time control of a system that includes time delay. Motivated by problems from Acoustic Positioning Research Inc., we consider the continual automated illumination of an object moving within a bounded domain, which requires object location prediction due to inherent mechanical and physical time lags associated with robotic lighting. Quality computational predictions demand high fidelity models for the coupled moving object signal and observation equipment pair. In our current problem, the signal represents the vector position, orientation, and velocity of a stage performer. Acoustic observations are formed by timing ultrasonic waves traveling from four perimeter speakers to a microphone attached to the performer. The goal is to schedule lighting movements that are coordinated with the performer by anticipating his/her future position based upon these observations using filtering theory. Particle system based methods have experienced rapid development and have become an essential technique of contemporary filtering strategies. Hitherto, researchers have largely focused on continuous state particle filters, ranging from traditional weighted particle filters to adaptive refining particle filters, readily able to perform path-space estimation and prediction. Herein, we compare the performance of a state-of-the-art refining particle filter to that of a novel discrete-space particle filter on the acoustic positioning problem. By discrete space particle filter we mean a Markov chain that counts particles in discretized cells of the signal state space in order to form an approximated unnormalized distribution of the signal state. For both filters mentioned above, we will examine issues like the mean time to localize a signal, the fidelity of filter estimates at various signal to noise ratios, computational costs, and the effect of signal

  9. Formant Centralization Ratio (FCR): A proposal for a new acoustic measure of dysarthric speech

    PubMed Central

    Sapir, Shimon; Ramig, Lorraine O.; Spielman, Jennifer L.; Fox, Cynthia

    2009-01-01

    Background and Aims The vowel space area (VSA) has been used as an acoustic metric of dysarthric speech, but with varying degrees of success. Here we test an alternative metric -- Formant centralization ratio (FCR) -- that is hypothesized to more effectively differentiate dysarthric from healthy speech and register treatment effects. Methods Speech recordings of 38 individuals with idiopathic Parkinson disease (IPD) and dysarthria (19 of whom received one month of intensive speech therapy (LSVT® LOUD)) and 14 healthy controls were acoustically analyzed. Vowels were extracted from short phrases. The same vowel-formant elements were used to construct the FCR, expressed as (F2u+F2𝔞+F1i+F1u)/(F2i+F1𝔞), the VSA, expressed as ABS((F1i*(F2𝔞–F2u)+F1𝔞*(F2u–F2i)+F1u*(F2i–F2𝔞))/2), a logarithmically scaled version of the VSA (LnVSA), and the F2i/F2u ratio. Results Unlike the VSA and the LnVSA, the FCR and F2i/F2u robustly differentiated dysarthric from healthy speech and were not gender-sensitive. All metrics effectively registered treatment effects and were strongly correlated with each other. Conclusions Albeit preliminary, the present findings indicate that the FCR is a sensitive, valid and reliable acoustic metric for distinguishing dysarthric from normal speech and for monitoring treatment effects, probably so because of reduced sensitivity to inter-speaker variability and enhanced sensitivity to vowel centralization. PMID:19948755

  10. Vibro-Acoustic Analysis of NASA's Space Shuttle Launch Pad 39A Flame Trench Wall

    NASA Technical Reports Server (NTRS)

    Margasahayam, Ravi N.

    2009-01-01

    A vital element to NASA's manned space flight launch operations is the Kennedy Space Center Launch Complex 39's launch pads A and B. Originally designed and constructed In the 1960s for the Saturn V rockets used for the Apollo missions, these pads were modified above grade to support Space Shuttle missions. But below grade, each of the pad's original walls (including a 42 feet deep, 58 feet wide, and 450 feet long tunnel designed to deflect flames and exhaust gases, the flame trench) remained unchanged. On May 31, 2008 during the launch of STS-124, over 3500 of the. 22000 interlocking refractory bricks that lined east wall of the flame trench, protecting the pad structure were liberated from pad 39A. The STS-124 launch anomaly spawned an agency-wide initiative to determine the failure root cause, to assess the impact of debris on vehicle and ground support equipment safety, and to prescribe corrective action. The investigation encompassed radar imaging, infrared video review, debris transport mechanism analysis using computational fluid dynamics, destructive testing, and non-destructive evaluation, including vibroacoustic analysis, in order to validate the corrective action. The primary focus of this paper is on the analytic approach, including static, modal, and vibro-acoustic analysis, required to certify the corrective action, and ensure Integrity and operational reliability for future launches. Due to the absence of instrumentation (including pressure transducers, acoustic pressure sensors, and accelerometers) in the flame trench, defining an accurate acoustic signature of the launch environment during shuttle main engine/solid rocket booster Ignition and vehicle ascent posed a significant challenge. Details of the analysis, including the derivation of launch environments, the finite element approach taken, and analysistest/ launch data correlation are discussed. Data obtained from the recent launch of STS-126 from Pad 39A was instrumental in validating the

  11. The Effect of Acoustic Disturbances on the Operation of the Space Shuttle Main Engine Fuel Flowmeter

    NASA Technical Reports Server (NTRS)

    Marcu, Bogdan; Szabo, Roland; Dorney, Dan; Zoladz, Tom

    2007-01-01

    The Space Shuttle Main Engine (SSME) uses a turbine fuel flowmeter (FFM) in its Low Pressure Fuel Duct (LPFD) to measure liquid hydrogen flowrates during engine operation. The flowmeter is required to provide accurate and robust measurements of flow rates ranging from 10000 to 18000 GPM in an environment contaminated by duct vibration and duct internal acoustic disturbances. Errors exceeding 0.5% can have a significant impact on engine operation and mission completion. The accuracy of each sensor is monitored during hot-fire engine tests on the ground. Flow meters which do not meet requirements are not flown. Among other parameters, the device is screened for a specific behavior in which a small shift in the flow rate reading is registered during a period in which the actual fuel flow as measured by a facility meter does not change. Such behavior has been observed over the years for specific builds of the FFM and must be avoided or limited in magnitude in flight. Various analyses of the recorded data have been made prior to this report in an effort to understand the cause of the phenomenon; however, no conclusive cause for the shift in the instrument behavior has been found. The present report proposes an explanation of the phenomenon based on interactions between acoustic pressure disturbances in the duct and the wakes produced by the FFM flow straightener. Physical insight into the effects of acoustic plane wave disturbances was obtained using a simple analytical model. Based on that model, a series of three-dimensional unsteady viscous flow computational fluid dynamics (CFD) simulations were performed using the MSFC PHANTOM turbomachinery code. The code was customized to allow the FFM rotor speed to change at every time step according to the instantaneous fluid forces on the rotor, that, in turn, are affected by acoustic plane pressure waves propagating through the device. The results of the simulations show the variation in the rotation rate of the flowmeter

  12. Reverberation impairs brainstem temporal representations of voiced vowel sounds: challenging "periodicity-tagged" segregation of competing speech in rooms.

    PubMed

    Sayles, Mark; Stasiak, Arkadiusz; Winter, Ian M

    2014-01-01

    The auditory system typically processes information from concurrently active sound sources (e.g., two voices speaking at once), in the presence of multiple delayed, attenuated and distorted sound-wave reflections (reverberation). Brainstem circuits help segregate these complex acoustic mixtures into "auditory objects." Psychophysical studies demonstrate a strong interaction between reverberation and fundamental-frequency (F0) modulation, leading to impaired segregation of competing vowels when segregation is on the basis of F0 differences. Neurophysiological studies of complex-sound segregation have concentrated on sounds with steady F0s, in anechoic environments. However, F0 modulation and reverberation are quasi-ubiquitous. We examine the ability of 129 single units in the ventral cochlear nucleus (VCN) of the anesthetized guinea pig to segregate the concurrent synthetic vowel sounds /a/ and /i/, based on temporal discharge patterns under closed-field conditions. We address the effects of added real-room reverberation, F0 modulation, and the interaction of these two factors, on brainstem neural segregation of voiced speech sounds. A firing-rate representation of single-vowels' spectral envelopes is robust to the combination of F0 modulation and reverberation: local firing-rate maxima and minima across the tonotopic array code vowel-formant structure. However, single-vowel F0-related periodicity information in shuffled inter-spike interval distributions is significantly degraded in the combined presence of reverberation and F0 modulation. Hence, segregation of double-vowels' spectral energy into two streams (corresponding to the two vowels), on the basis of temporal discharge patterns, is impaired by reverberation; specifically when F0 is modulated. All unit types (primary-like, chopper, onset) are similarly affected. These results offer neurophysiological insights to perceptual organization of complex acoustic scenes under realistically challenging listening

  13. Reverberation impairs brainstem temporal representations of voiced vowel sounds: challenging “periodicity-tagged” segregation of competing speech in rooms

    PubMed Central

    Sayles, Mark; Stasiak, Arkadiusz; Winter, Ian M.

    2015-01-01

    The auditory system typically processes information from concurrently active sound sources (e.g., two voices speaking at once), in the presence of multiple delayed, attenuated and distorted sound-wave reflections (reverberation). Brainstem circuits help segregate these complex acoustic mixtures into “auditory objects.” Psychophysical studies demonstrate a strong interaction between reverberation and fundamental-frequency (F0) modulation, leading to impaired segregation of competing vowels when segregation is on the basis of F0 differences. Neurophysiological studies of complex-sound segregation have concentrated on sounds with steady F0s, in anechoic environments. However, F0 modulation and reverberation are quasi-ubiquitous. We examine the ability of 129 single units in the ventral cochlear nucleus (VCN) of the anesthetized guinea pig to segregate the concurrent synthetic vowel sounds /a/ and /i/, based on temporal discharge patterns under closed-field conditions. We address the effects of added real-room reverberation, F0 modulation, and the interaction of these two factors, on brainstem neural segregation of voiced speech sounds. A firing-rate representation of single-vowels' spectral envelopes is robust to the combination of F0 modulation and reverberation: local firing-rate maxima and minima across the tonotopic array code vowel-formant structure. However, single-vowel F0-related periodicity information in shuffled inter-spike interval distributions is significantly degraded in the combined presence of reverberation and F0 modulation. Hence, segregation of double-vowels' spectral energy into two streams (corresponding to the two vowels), on the basis of temporal discharge patterns, is impaired by reverberation; specifically when F0 is modulated. All unit types (primary-like, chopper, onset) are similarly affected. These results offer neurophysiological insights to perceptual organization of complex acoustic scenes under realistically challenging listening

  14. Comparative evaluation of Space Transportation System (STS)-3 flight and acoustic test random vibration response of the OSS-1 payload

    NASA Technical Reports Server (NTRS)

    On, F. J.

    1983-01-01

    A comparative evaluation of the Space Transportation System (STS)-3 flight and acoustic test random vibration response of the Office of Space Science-1 (OSS-1) payload is presented. The results provide insight into the characteristics of vibroacoustic response of pallet payload components in the payload bay during STS flights.

  15. Enhancement of temporal periodicity cues in cochlear implants: Effects on prosodic perception and vowel identification

    NASA Astrophysics Data System (ADS)

    Green, Tim; Faulkner, Andrew; Rosen, Stuart; Macherey, Olivier

    2005-07-01

    Standard continuous interleaved sampling processing, and a modified processing strategy designed to enhance temporal cues to voice pitch, were compared on tests of intonation perception, and vowel perception, both in implant users and in acoustic simulations. In standard processing, 400 Hz low-pass envelopes modulated either pulse trains (implant users) or noise carriers (simulations). In the modified strategy, slow-rate envelope modulations, which convey dynamic spectral variation crucial for speech understanding, were extracted by low-pass filtering (32 Hz). In addition, during voiced speech, higher-rate temporal modulation in each channel was provided by 100% amplitude-modulation by a sawtooth-like wave form whose periodicity followed the fundamental frequency (F0) of the input. Channel levels were determined by the product of the lower- and higher-rate modulation components. Both in acoustic simulations and in implant users, the ability to use intonation information to identify sentences as question or statement was significantly better with modified processing. However, while there was no difference in vowel recognition in the acoustic simulation, implant users performed worse with modified processing both in vowel recognition and in formant frequency discrimination. It appears that, while enhancing pitch perception, modified processing harmed the transmission of spectral information.

  16. Calibration of International Space Station (ISS) Node 1 Vibro-Acoustic Model

    NASA Technical Reports Server (NTRS)

    Zhang, Weiguo; Raveendra, Ravi

    2014-01-01

    Reported here is the ability of utilizing the Energy Finite Element Method (E-FEM) to predict the vibro-acoustic sound fields within the International Space Station (ISS) Node 1 and to compare the results with actual measurements of leak sounds made by a one atmosphere to vacuum leak through a small hole in the pressure wall of the Node 1 STA module during its period of storage at Stennis Space Center (SSC). While the E-FEM method represents a reverberant sound field calculation, of importance to this application is the requirement to also handle the direct field effect of the sound generation. It was also important to be able to compute the sound fields in the ultrasonic frequency range. This report demonstrates the capability of this technology as applied to this type of application.

  17. Acoustic omni meta-atom for decoupled access to all octants of a wave parameter space

    NASA Astrophysics Data System (ADS)

    Koo, Sukmo; Cho, Choonlae; Jeong, Jun-Ho; Park, Namkyoo

    2016-09-01

    The common behaviour of a wave is determined by wave parameters of its medium, which are generally associated with the characteristic oscillations of its corresponding elementary particles. In the context of metamaterials, the decoupled excitation of these fundamental oscillations would provide an ideal platform for top-down and reconfigurable access to the entire constitutive parameter space; however, this has remained as a conceivable problem that must be accomplished, after being pointed out by Pendry. Here by focusing on acoustic metamaterials, we achieve the decoupling of density ρ, modulus B-1 and bianisotropy ξ, by separating the paths of particle momentum to conform to the characteristic oscillations of each macroscopic wave parameter. Independent access to all octants of wave parameter space (ρ, B-1, ξ)=(+/-,+/-,+/-) is thus realized using a single platform that we call an omni meta-atom; as a building block that achieves top-down access to the target properties of metamaterials.

  18. Acoustic omni meta-atom for decoupled access to all octants of a wave parameter space

    PubMed Central

    Koo, Sukmo; Cho, Choonlae; Jeong, Jun-ho; Park, Namkyoo

    2016-01-01

    The common behaviour of a wave is determined by wave parameters of its medium, which are generally associated with the characteristic oscillations of its corresponding elementary particles. In the context of metamaterials, the decoupled excitation of these fundamental oscillations would provide an ideal platform for top–down and reconfigurable access to the entire constitutive parameter space; however, this has remained as a conceivable problem that must be accomplished, after being pointed out by Pendry. Here by focusing on acoustic metamaterials, we achieve the decoupling of density ρ, modulus B−1 and bianisotropy ξ, by separating the paths of particle momentum to conform to the characteristic oscillations of each macroscopic wave parameter. Independent access to all octants of wave parameter space (ρ, B−1, ξ)=(+/−,+/−,+/−) is thus realized using a single platform that we call an omni meta-atom; as a building block that achieves top–down access to the target properties of metamaterials. PMID:27687689

  19. Multiple mechanisms for recency with vowels and consonants.

    PubMed

    Battacchi, M W; Pelamatti, G M; Umiltà, C

    1989-05-01

    Three experiments were conducted to test the hypothesis that the difference in recency effect between vowel-contrasting and stop-contrasting lists of syllables in immediate ordered recall can be explained by item discriminability and regular short-term memory mechanisms, without any recourse to echoic memory or precategorical acoustic storage (PAS). In Experiment 1, the short-term memory mechanisms were manipulated by reducing amount of output interference and length of retention interval. The partial-report technique was used. The most important finding was the usual final-position recency effect (difference in recall between the fifth and sixth serial positions) for the vowel lists but not for the stop lists, regardless of the type of report. Thus the PAS theory could not be rejected. In Experiments 2 and 3, the last item was differentiated from the other items of the list, either by lengthening the interstimulus interval between the last and the next-to-last (Experiment 2) or by increasing the intensity of the last item (Experiment 3). In both cases, an increase of the final-position recency effect was found even for stop lists. Since a drop in recall errors was also obtained for the fourth item when its intensity was increased (von Restorff effect), this final-position recency effect for stop lists is likely to be due to item discriminability, and not to echoic memory for the last item. Item discriminability appeared to be the critical factor.

  20. Specific features of vowel-like signals of white whales

    NASA Astrophysics Data System (ADS)

    Bel'Kovich, V. M.; Kreichi, S. A.

    2004-05-01

    The set of acoustic signals of White-Sea white whales comprises about 70 types of signals. Six of them occur most often and constitute 75% of the total number of signals produced by these animals. According to behavioral reactions, white whales distinguish each other by acoustic signals, which is also typical of other animal species and humans. To investigate this phenomenon, signals perceived as vowel-like sounds of speech, including sounds perceived as a “bleat,” were chosen A sample of 480 signals recorded in June and July, 2000, in the White Sea within a reproductive assemblage of white whales near the Large Solovetskii Island was studied. Signals were recorded on a digital data carrier (a SONY minidisk) in the frequency range of 0.06 20 kHz. The purpose of the study was to reveal the perceptive and acoustic features specific to individual animals. The study was carried out using the methods of structural analysis of vocal speech that are employed in lingual criminalistics to identify a speaking person. It was demonstrated that this approach allows one to group the signals by coincident perceptive and acoustic parameters with assigning individual attributes to single parameters. This provided an opportunity to separate conditionally about 40 different sources of acoustic signals according to the totality of coincidences, which corresponded to the number of white whales observed visually. Thus, the application of this method proves to be very promising for the acoustic identification of white whales and other marine mammals, this possibility being very important for biology.

  1. Minke whale song, spacing, and acoustic communication on the Great Barrier Reef, Australia

    NASA Astrophysics Data System (ADS)

    Gedamke, Jason

    An inquisitive population of minke whale (Balaenoptera acutorostrata ) that concentrates on the Great Barrier Reef during its suspected breeding season offered a unique opportunity to conduct a multi-faceted study of a little-known Balaenopteran species' acoustic behavior. Chapter one investigates whether the minke whale is the source of an unusual, complex, and stereotyped sound recorded, the "star-wars" vocalization. A hydrophone array was towed from a vessel to record sounds from circling whales for subsequent localization of sound sources. These acoustic locations were matched with shipboard and in-water observations of the minke whale, demonstrating the minke whale was the source of this unusual sound. Spectral and temporal features of this sound and the source levels at which it is produced are described. The repetitive "star-wars" vocalization appears similar to the songs of other whale species and has characteristics consistent with reproductive advertisement displays. Chapter two investigates whether song (i.e. the "star-wars" vocalization) has a spacing function through passive monitoring of singer spatial patterns with a moored five-sonobuoy array. Active song playback experiments to singers were also conducted to further test song function. This study demonstrated that singers naturally maintain spatial separations between them through a nearest-neighbor analysis and animated tracks of singer movements. In response to active song playbacks, singers generally moved away and repeated song more quickly suggesting that song repetition interval may help regulate spatial interaction and singer separation. These results further indicate the Great Barrier Reef may be an important reproductive habitat for this species. Chapter three investigates whether song is part of a potentially graded repertoire of acoustic signals. Utilizing both vessel-based recordings and remote recordings from the sonobuoy array, temporal and spectral features, source levels, and

  2. The right ear advantage revisited: speech lateralisation in dichotic listening using consonant-vowel and vowel-consonant syllables.

    PubMed

    Sætrevik, Bjørn

    2012-01-01

    The dichotic listening task is typically administered by presenting a consonant-vowel (CV) syllable to each ear and asking the participant to report the syllable heard most clearly. The results tend to show more reports of the right ear syllable than of the left ear syllable, an effect called the right ear advantage (REA). The REA is assumed to be due to the crossing over of auditory fibres and the processing of language stimuli being lateralised to left temporal areas. However, the tendency for most dichotic listening experiments to use only CV syllable stimuli limits the extent to which the conclusions can be generalised to also apply to other speech phonemes. The current study re-examines the REA in dichotic listening by using both CV and vowel-consonant (VC) syllables and combinations thereof. Results showed a replication of the REA response pattern for both CV and VC syllables, thus indicating that the general assumption of left-side localisation of processing can be applied for both types of stimuli. Further, on trials where a CV is presented in one ear and a VC is presented in the other ear, the CV is selected more often than the VC, indicating that these phonemes have an acoustic or processing advantage.

  3. Parsing the role of consonants versus vowels in the classic Takete-Maluma phenomenon.

    PubMed

    Nielsen, Alan K S; Rendall, Drew

    2013-06-01

    Wolfgang Köhler (1929, Gestalt psychology, New York, NY: Liveright) famously reported a bias in people's choice of nonsense words as labels for novel objects, pointing to possible naïve expectations about language structure. Two accounts have been offered to explain this bias, one focusing on the visuomotor effects of different vowel forms and the other focusing on variation in the acoustic structure and perceptual quality of different consonants. To date, evidence in support of both effects is mixed. Moreover, the veracity of either effect has often been doubted due to perceived limitations in methodologies and stimulus materials. A novel word-construction experiment is presented to test both proposed effects using randomized word- and image-generation techniques to address previous methodological concerns. Results show that participants are sensitive to both vowel and consonant content, constructing novel words of relatively sonorant consonants and rounded vowels to label curved object images, and of relatively plosive consonants and nonrounded vowels to label jagged object images. Results point to additional influences on word construction potentially related to the articulatory affordances or constraints accompanying different word forms. PMID:23205509

  4. English vowel identification in quiet and noise: effects of listeners' native language background

    PubMed Central

    Jin, Su-Hyun; Liu, Chang

    2014-01-01

    Purpose: To investigate the effect of listener's native language (L1) and the types of noise on English vowel identification in noise. Method: Identification of 12 English vowels was measured in quiet and in long-term speech-shaped noise and multi-talker babble (MTB) noise for English- (EN), Chinese- (CN) and Korean-native (KN) listeners at various signal-to-noise ratios (SNRs). Results: Compared to non-native listeners, EN listeners performed significantly better in quiet and in noise. Vowel identification in long-term speech-shaped noise and in MTB noise was similar between CN and KN listeners. This is different from our previous study in which KN listeners performed better than CN listeners in English sentence recognition in MTB noise. Discussion: Results from the current study suggest that depending on speech materials, the effect of non-native listeners' L1 on speech perception in noise may be different. That is, in the perception of speech materials with little linguistic cues like isolated vowels, the characteristics of non-native listener's native language may not play a significant role. On the other hand, in the perception of running speech in which listeners need to use more linguistic cues (e.g., acoustic-phonetic, semantic, and prosodic cues), the non-native listener's native language background might result in a different masking effect. PMID:25400538

  5. Processing interactions between phonology and melody: vowels sing but consonants speak.

    PubMed

    Kolinsky, Régine; Lidji, Pascale; Peretz, Isabelle; Besson, Mireille; Morais, José

    2009-07-01

    The aim of this study was to determine if two dimensions of song, the phonological part of lyrics and the melodic part of tunes, are processed in an independent or integrated way. In a series of five experiments, musically untrained participants classified bi-syllabic nonwords sung on two-tone melodic intervals. Their response had to be based on pitch contour, on nonword identity, or on the combination of pitch and nonword. When participants had to ignore irrelevant variations of the non-attended dimension, patterns of interference and facilitation allowed us to specify the processing interactions between dimensions. Results showed that consonants are processed more independently from melodic information than vowels are (Experiments 1-4). This difference between consonants and vowels was neither related to the sonority of the phoneme (Experiment 3), nor to the acoustical correlates between vowel quality and pitch height (Experiment 5). The implication of these results for our understanding of the functional relationships between musical and linguistic systems is discussed in light of the different evolutionary origins and linguistic functions of consonants and vowels.

  6. The discrimination of baboon grunt calls and human vowel sounds by baboons

    NASA Astrophysics Data System (ADS)

    Hienz, Robert D.; Jones, April M.; Weerts, Elise M.

    2004-09-01

    The ability of baboons to discriminate changes in the formant structures of a synthetic baboon grunt call and an acoustically similar human vowel (/eh/) was examined to determine how comparable baboons are to humans in discriminating small changes in vowel sounds, and whether or not any species-specific advantage in discriminability might exist when baboons discriminate their own vocalizations. Baboons were trained to press and hold down a lever to produce a pulsed train of a standard sound (e.g., /eh/ or a baboon grunt call), and to release the lever only when a variant of the sound occurred. Synthetic variants of each sound had the same first and third through fifth formants (F1 and F3-5), but varied in the location of the second formant (F2). Thresholds for F2 frequency changes were 55 and 67 Hz for the grunt and vowel stimuli, respectively, and were not statistically different from one another. Baboons discriminated changes in vowel formant structures comparable to those discriminated by humans. No distinct advantages in discrimination performances were observed when the baboons discriminated these synthetic grunt vocalizations.

  7. Parsing the role of consonants versus vowels in the classic Takete-Maluma phenomenon.

    PubMed

    Nielsen, Alan K S; Rendall, Drew

    2013-06-01

    Wolfgang Köhler (1929, Gestalt psychology, New York, NY: Liveright) famously reported a bias in people's choice of nonsense words as labels for novel objects, pointing to possible naïve expectations about language structure. Two accounts have been offered to explain this bias, one focusing on the visuomotor effects of different vowel forms and the other focusing on variation in the acoustic structure and perceptual quality of different consonants. To date, evidence in support of both effects is mixed. Moreover, the veracity of either effect has often been doubted due to perceived limitations in methodologies and stimulus materials. A novel word-construction experiment is presented to test both proposed effects using randomized word- and image-generation techniques to address previous methodological concerns. Results show that participants are sensitive to both vowel and consonant content, constructing novel words of relatively sonorant consonants and rounded vowels to label curved object images, and of relatively plosive consonants and nonrounded vowels to label jagged object images. Results point to additional influences on word construction potentially related to the articulatory affordances or constraints accompanying different word forms.

  8. Acoustic puncture assist device versus loss of resistance technique for epidural space identification

    PubMed Central

    Mittal, Amit Kumar; Goel, Nitesh; Chowdhury, Itee; Shah, Shagun Bhatia; Singh, Brijesh Pratap; Jakhar, Pradeep

    2016-01-01

    Background and Aims: The conventional techniques of epidural space (EDS) identification based on loss of resistance (LOR) have a higher chance of complications, patchy analgesia and epidural failure, which can be minimised by objective confirmation of space before catheter placement. Acoustic puncture assist device (APAD) technique objectively confirms EDS, thus enhancing success, with lesser complications. This study was planned with the objective to evaluate the APAD technique and compare it to LOR technique for EDS identification and its correlation with ultrasound guided EDS depth. Methods: In this prospective study, the lumbar vertebral spaces were scanned by the ultrasound for measuring depth of the EDS and later correlated with procedural depth measured by either of the technique (APAD or LOR). The data were subjected to descriptive statistics; the concordance correlation coefficient and Bland-Altman analysis with 95% confidence limits. Results: Acoustic dip in pitch and descent in pressure tracing on EDS localisation was observed among the patients of APAD group. Analysis of concordance correlation between the ultrasonography (USG) depth and APAD or LOR depth was significant (r ≥ 0.97 in both groups). Bland-Altman analysis revealed a mean difference of 0.171cm in group APAD and 0.154 cm in group LOR. The 95% limits of agreement for the difference between the two measurements were − 0.569 and 0.226 cm in APAD and − 0.530 to 0.222 cm in LOR group. Conclusion: We found APAD to be a precise tool for objective localisation of the EDS, co-relating well with the pre-procedural USG depth of EDS. PMID:27212720

  9. Intelligibility of American English vowels and consonants spoken by international students in the United States.

    PubMed

    Jin, Su-Hyun; Liu, Chang

    2014-04-01

    PURPOSE The purpose of this study was to examine the intelligibility of English consonants and vowels produced by Chinese-native (CN), and Korean-native (KN) students enrolled in American universities. METHOD 16 English-native (EN), 32 CN, and 32 KN speakers participated in this study. The intelligibility of 16 American English consonants and 16 vowels spoken by native and nonnative speakers of English was evaluated by EN listeners. All nonnative speakers also completed a survey of their language backgrounds. RESULTS Although the intelligibility of consonants and diphthongs for nonnative speakers was comparable to that of native speakers, the intelligibility of monophthongs was significantly lower for CN and KN speakers than for EN speakers. Sociolinguistic factors such as the age of arrival in the United States and daily use of English, as well as a linguistic factor, difference in vowel space between native (L1) and nonnative (L2) language, partially contributed to vowel intelligibility for CN and KN groups. There was no significant correlation between the length of U.S. residency and phoneme intelligibility. CONCLUSION Results indicated that the major difficulty in phonemic production in English for Chinese and Korean speakers is with vowels rather than consonants. This might be useful for developing training methods to improve English intelligibility for foreign students in the United States.

  10. An investigation of acoustic noise requirements for the Space Station centrifuge facility

    NASA Technical Reports Server (NTRS)

    Castellano, Timothy

    1994-01-01

    Acoustic noise emissions from the Space Station Freedom (SSF) centrifuge facility hardware represent a potential technical and programmatic risk to the project. The SSF program requires that no payload exceed a Noise Criterion 40 (NC-40) noise contour in any octave band between 63 Hz and 8 kHz as measured 2 feet from the equipment item. Past experience with life science experiment hardware indicates that this requirement will be difficult to meet. The crew has found noise levels on Spacelab flights to be unacceptably high. Many past Ames Spacelab life science payloads have required waivers because of excessive noise. The objectives of this study were (1) to develop an understanding of acoustic measurement theory, instruments, and technique, and (2) to characterize the noise emission of analogous Facility components and previously flown flight hardware. Test results from existing hardware were reviewed and analyzed. Measurements of the spectral and intensity characteristics of fans and other rotating machinery were performed. The literature was reviewed and contacts were made with NASA and industry organizations concerned with or performing research on noise control.

  11. Frequency-space prediction filtering for acoustic clutter and random noise attenuation in ultrasound imaging

    NASA Astrophysics Data System (ADS)

    Shin, Junseob; Huang, Lianjie

    2016-04-01

    Frequency-space prediction filtering (FXPF), also known as FX deconvolution, is a technique originally developed for random noise attenuation in seismic imaging. FXPF attempts to reduce random noise in seismic data by modeling only real signals that appear as linear or quasilinear events in the aperture domain. In medical ultrasound imaging, channel radio frequency (RF) signals from the main lobe appear as horizontal events after receive delays are applied while acoustic clutter signals from off-axis scatterers and electronic noise do not. Therefore, FXPF is suitable for preserving only the main-lobe signals and attenuating the unwanted contributions from clutter and random noise in medical ultrasound imaging. We adapt FXPF to ultrasound imaging, and evaluate its performance using simulated data sets from a point target and an anechoic cyst. Our simulation results show that using only 5 iterations of FXPF achieves contrast-to-noise ratio (CNR) improvements of 67 % in a simulated noise-free anechoic cyst and 228 % in a simulated anechoic cyst contaminated with random noise of 15 dB signal-to-noise ratio (SNR). Our findings suggest that ultrasound imaging with FXPF attenuates contributions from both acoustic clutter and random noise and therefore, FXPF has great potential to improve ultrasound image contrast for better visualization of important anatomical structures and detection of diseased conditions.

  12. Perturbation and Nonlinear Dynamic Analysis of Acoustic Phonatory Signal in Parkinsonian Patients Receiving Deep Brain Stimulation

    ERIC Educational Resources Information Center

    Lee, Victoria S.; Zhou, Xiao Ping; Rahn, Douglas A., III; Wang, Emily Q.; Jiang, Jack J.

    2008-01-01

    Nineteen PD patients who received deep brain stimulation (DBS), 10 non-surgical (control) PD patients, and 11 non-pathologic age- and gender-matched subjects performed sustained vowel phonations. The following acoustic measures were obtained on the sustained vowel phonations: correlation dimension (D[subscript 2]), percent jitter, percent shimmer,…

  13. The effect of magnetic field-free space on the acoustic behavior of budgerigars (Melopsittacus undulafus)

    NASA Astrophysics Data System (ADS)

    Jiang, Jin-Chang; Jin, Hai-Qiang; Lin, Yun-Fang; Chen, Hao; Yang, Xin-Yu; Zeng, Xiao-Ping; Zhou, Xun

    1998-07-01

    The effect of magnetic field-free space (MFFS) on the acoustic behavior of budgerigar (Melopsittacus undulafus) is obvious. The daily frequency of their cries in uniform MFFS (UMFFS) and non-uniform MFFS (NMFFS) decreases by 44.7±10.0% as compared with that in the geomagnetic field (GMF) on the average. The occupation rate of protesting cries (R op) in NMFFS decreases by 8.5% 20.3% as compared with that in GMF on the average and shows an adaptability variation. In 75% of the observation days, the R op in UMFFS increases by 16.2% 23.3% as compared with that in GMF. As for the effect of MFFS on the rhythmic habits of budgerigars, only the ending time of crying is affected to certain extent, 67 minutes earlier than in GMF on the average.

  14. Baryon acoustic oscillations in 2D: Modeling redshift-space power spectrum from perturbation theory

    NASA Astrophysics Data System (ADS)

    Taruya, Atsushi; Nishimichi, Takahiro; Saito, Shun

    2010-09-01

    We present an improved prescription for the matter power spectrum in redshift space taking proper account of both nonlinear gravitational clustering and redshift distortion, which are of particular importance for accurately modeling baryon acoustic oscillations (BAOs). Contrary to the models of redshift distortion phenomenologically introduced but frequently used in the literature, the new model includes the corrections arising from the nonlinear coupling between the density and velocity fields associated with two competitive effects of redshift distortion, i.e., Kaiser and Finger-of-God effects. Based on the improved treatment of perturbation theory for gravitational clustering, we compare our model predictions with the monopole and quadrupole power spectra of N-body simulations, and an excellent agreement is achieved over the scales of BAOs. Potential impacts on constraining dark energy and modified gravity from the redshift-space power spectrum are also investigated based on the Fisher-matrix formalism, particularly focusing on the measurements of the Hubble parameter, angular diameter distance, and growth rate for structure formation. We find that the existing phenomenological models of redshift distortion produce a systematic error on measurements of the angular diameter distance and Hubble parameter by 1%-2% , and the growth-rate parameter by ˜5%, which would become non-negligible for future galaxy surveys. Correctly modeling redshift distortion is thus essential, and the new prescription for the redshift-space power spectrum including the nonlinear corrections can be used as an accurate theoretical template for anisotropic BAOs.

  15. Acoustic puncture assist device: A novel technique to identify the epidural space

    PubMed Central

    Al-Mokaddam, MA; Al-Harbi, MK; El-Jandali, ST; Al-Zahrani, TA

    2016-01-01

    Background: Acoustic puncture assist device (APAD) is designed to detect and signal the loss of resistance during the epidural procedure. We aimed to evaluate this device in terms of successful identification of the epidural space and the incidence of accidental dural puncture. Patients and Methods: Following Institutional Review Board approval and written informed consent obtained from all patients, 200 adult patients (107 males) American Society of Anesthesiologists I-III who underwent lower limb orthopedic surgery under lumbar epidural anesthesia using APAD were enrolled in the study. APAD system was connected to the epidural needle using normal saline prefilled extension tube. Numbers of successful epidural attempts and accidental dural tap were documented. Results: The mean values of the depth of epidural space and the time to perform epidural puncture were 5.8 ± 1.0 cm and 3.3 ± 1.4 min, respectively. In 63% of patients, epidural puncture was successful from the first attempt and in 1% it was successful from the fourth attempt. Epidural anesthesia by APAD was successful in 198 cases (99 %). Dural tap occurred in 2 cases (1%). Conclusions: Using APAD, the success of identifying the epidural space was high and reliable. PMID:27051369

  16. Baryon acoustic oscillations in 2D: Modeling redshift-space power spectrum from perturbation theory

    SciTech Connect

    Taruya, Atsushi; Nishimichi, Takahiro; Saito, Shun

    2010-09-15

    We present an improved prescription for the matter power spectrum in redshift space taking proper account of both nonlinear gravitational clustering and redshift distortion, which are of particular importance for accurately modeling baryon acoustic oscillations (BAOs). Contrary to the models of redshift distortion phenomenologically introduced but frequently used in the literature, the new model includes the corrections arising from the nonlinear coupling between the density and velocity fields associated with two competitive effects of redshift distortion, i.e., Kaiser and Finger-of-God effects. Based on the improved treatment of perturbation theory for gravitational clustering, we compare our model predictions with the monopole and quadrupole power spectra of N-body simulations, and an excellent agreement is achieved over the scales of BAOs. Potential impacts on constraining dark energy and modified gravity from the redshift-space power spectrum are also investigated based on the Fisher-matrix formalism, particularly focusing on the measurements of the Hubble parameter, angular diameter distance, and growth rate for structure formation. We find that the existing phenomenological models of redshift distortion produce a systematic error on measurements of the angular diameter distance and Hubble parameter by 1%-2%, and the growth-rate parameter by {approx}5%, which would become non-negligible for future galaxy surveys. Correctly modeling redshift distortion is thus essential, and the new prescription for the redshift-space power spectrum including the nonlinear corrections can be used as an accurate theoretical template for anisotropic BAOs.

  17. Space Shuttle Orbiter Main Engine Ignition Acoustic Pressure Loads Issue: Recent Actions to Install Wireless Instrumentation on STS-129

    NASA Technical Reports Server (NTRS)

    Wells, Nathan; Studor, George

    2009-01-01

    This slide presentation reviews the development and construction of the wireless acoustic instruments surrounding the space shuttle's main engines in preparation for STS-129. The presentation also includes information on end-of-life processing and the mounting procedure for the devices.

  18. Study for Identification of Beneficial Uses of Space (BUS). Volume 2: Technical report. Book 4: Development and business analysis of space processed surface acoustic wave devices

    NASA Technical Reports Server (NTRS)

    1975-01-01

    Preliminary development plans, analysis of required R and D and production resources, the costs of such resources, and, finally, the potential profitability of a commercial space processing opportunity for the production of very high frequency surface acoustic wave devices are presented.

  19. Palatalization and Intrinsic Prosodic Vowel Features in Russian

    ERIC Educational Resources Information Center

    Ordin, Mikhail

    2011-01-01

    The presented study is aimed at investigating the interaction of palatalization and intrinsic prosodic features of the vowel in CVC (consonant+vowel+consonant) syllables in Russian. The universal nature of intrinsic prosodic vowel features was confirmed with the data from the Russian language. It was found that palatalization of the consonants…

  20. Audiovisual Perception of Congruent and Incongruent Dutch Front Vowels

    ERIC Educational Resources Information Center

    Valkenier, Bea; Duyne, Jurriaan Y.; Andringa, Tjeerd C.; Baskent, Deniz

    2012-01-01

    Purpose: Auditory perception of vowels in background noise is enhanced when combined with visually perceived speech features. The objective of this study was to investigate whether the influence of visual cues on vowel perception extends to incongruent vowels, in a manner similar to the McGurk effect observed with consonants. Method:…

  1. The Role of Consonant/Vowel Organization in Perceptual Discrimination

    ERIC Educational Resources Information Center

    Chetail, Fabienne; Drabs, Virginie; Content, Alain

    2014-01-01

    According to a recent hypothesis, the CV pattern (i.e., the arrangement of consonant and vowel letters) constrains the mental representation of letter strings, with each vowel or vowel cluster being the core of a unit. Six experiments with the same/different task were conducted to test whether this structure is extracted prelexically. In the…

  2. Factors Related to the Pronunciation of Vowel Clusters.

    ERIC Educational Resources Information Center

    Johnson, Dale D.

    This research report examines the pronunciation that children give to synthetic words containing vowel-cluster spellings and analyzes the observed pronunciations in relation to common English words containing the same vowel clusters. The pronunciations associated with vowel-cluster spellings are among the most unpredictable letter-sound…

  3. Reading in Thai: The Case of Misaligned Vowels

    ERIC Educational Resources Information Center

    Winskel, Heather

    2009-01-01

    Thai has its own distinctive alphabetic script with syllabic characteristics as it has implicit vowels for some consonants. Consonants are written in a linear order, but vowels can be written non-linearly above, below or to either side of the consonant. Of particular interest to the current study are that vowels can precede the consonant in…

  4. Optimizing cochlear implant frequency boundary tables for vowel perception: A computer simulation

    NASA Astrophysics Data System (ADS)

    Fourakis, Marios S.; Hawks, John W.; Schwager, Amy

    2001-05-01

    For cochlear implants, the assignment of frequency bands to electrodes is a variable parameter that determines what region of the acoustic spectrum will be represented by each electrode's output. Technology will soon allow for considerable flexibility in programming this parameter. In a first attempt to optimize these assignments for vowel perception, a computer program was written to evaluate different assignment schemes for categorization accuracy based strictly on the frequency values of the first two or three formants. Databases [J. Hillenbrand et al., J. Acoust. Soc. Am. 97, 3099-3111 (1995)] of formant measurements from American English vowels as uttered by men, women, and children were used. For this simulation, it was assumed that each formant frequency was associated with only the frequency band its center frequency fell within. Each pattern of frequency bands was assigned a vowel category identification based on the plurality of tokens whose intended identification category fell within that pattern. A range of frequency scaling schemes for 19 and 20 electrode arrays was evaluated, with the best of these fine tuned for minimum error. The results indicate that manufacturer's default assignments categorize reasonably well, but a bark-scaled scheme yielded the best unmodified classifications.

  5. An Amplitude-Based Estimation Method for International Space Station (ISS) Leak Detection and Localization Using Acoustic Sensor Networks

    NASA Technical Reports Server (NTRS)

    Tian, Jialin; Madaras, Eric I.

    2009-01-01

    The development of a robust and efficient leak detection and localization system within a space station environment presents a unique challenge. A plausible approach includes the implementation of an acoustic sensor network system that can successfully detect the presence of a leak and determine the location of the leak source. Traditional acoustic detection and localization schemes rely on the phase and amplitude information collected by the sensor array system. Furthermore, the acoustic source signals are assumed to be airborne and far-field. Likewise, there are similar applications in sonar. In solids, there are specialized methods for locating events that are used in geology and in acoustic emission testing that involve sensor arrays and depend on a discernable phase front to the received signal. These methods are ineffective if applied to a sensor detection system within the space station environment. In the case of acoustic signal location, there are significant baffling and structural impediments to the sound path and the source could be in the near-field of a sensor in this particular setting.

  6. Embedded Vowels: Remedying the Problems Arising out of Embedded Vowels in the English Writings of Arab Learners

    ERIC Educational Resources Information Center

    Khan, Mohamed Fazlulla

    2013-01-01

    L1 habits often tend to interfere with the process of learning a second language. The vowel habits of Arab learners of English are one such interference. Arabic orthography is such that certain vowels indicated by diacritics are often omitted, since an experienced reader of Arabic knows, by habit, the exact vowel sound in each phonetic…

  7. Learning Vowel Categories from Maternal Speech in Gurindji Kriol

    ERIC Educational Resources Information Center

    Jones, Caroline; Meakins, Felicity; Muawiyath, Shujau

    2012-01-01

    Distributional learning is a proposal for how infants might learn early speech sound categories from acoustic input before they know many words. When categories in the input differ greatly in relative frequency and overlap in acoustic space, research in bilingual development suggests that this affects the course of development. In the present…

  8. Calibration of International Space Station (ISS) Node 1 Vibro-Acoustic Model-Report 2

    NASA Technical Reports Server (NTRS)

    Zhang, Weiguo; Raveendra, Ravi

    2014-01-01

    Reported here is the capability of the Energy Finite Element Method (E-FEM) to predict the vibro-acoustic sound fields within the International Space Station (ISS) Node 1 and to compare the results with simulated leak sounds. A series of electronically generated structural ultrasonic noise sources were created in the pressure wall to emulate leak signals at different locations of the Node 1 STA module during its period of storage at Stennis Space Center (SSC). The exact sound source profiles created within the pressure wall at the source were unknown, but were estimated from the closest sensor measurement. The E-FEM method represents a reverberant sound field calculation, and of importance to this application is the requirement to correctly handle the direct field effect of the sound generation. It was also important to be able to compute the sound energy fields in the ultrasonic frequency range. This report demonstrates the capability of this technology as applied to this type of application.

  9. Online Damage Detection on Metal and Composite Space Structures by Active and Passive Acoustic Methods

    NASA Astrophysics Data System (ADS)

    Scheerer, M.; Cardone, T.; Rapisarda, A.; Ottaviano, S.; Ftancesconi, D.

    2012-07-01

    In the frame of ESA funded programme Future Launcher Preparatory Programme Period 1 “Preparatory Activities on M&S”, Aerospace & Advanced Composites and Thales Alenia Space-Italia, have conceived and tested a structural health monitoring approach based on integrated Acoustic Emission - Active Ultrasound Damage Identification. The monitoring methods implemented in the study are both passive and active methods and the purpose is to cover large areas with a sufficient damage size detection capability. Two representative space sub-structures have been built and tested: a composite overwrapped pressure vessel (COPV) and a curved, stiffened Al-Li panel. In each structure, typical critical damages have been introduced: delaminations caused by impacts in the COPV and a crack in the stiffener of the Al-Li panel which was grown during a fatigue test campaign. The location and severity of both types of damages have been successfully assessed online using two commercially available systems: one 6 channel AE system from Vallen and one 64 channel AU system from Acellent.

  10. Control of an experiment to measure acoustic noise in the space shuttle

    NASA Astrophysics Data System (ADS)

    Cameron, Charles B.

    1989-06-01

    The potential use of a general-purpose controller to measure acoustic vibration autonomously in the Space Shuttle Cargo Bay during launch is described. The experimental package will be housed in a Shuttle Get Away Special (GAS) canister. The control functions were implemented with software written largely in the C programming language. An IBM MS DOS computer and C cross-compiler were used to generate Z-80 assembly language code, assemble and link this code, and then transfer it to EPROM for use in the experiment's controller. The software is written in a modular fashion to permit adapting it easily to other applications. The software combines the experimental control functions with a menu-driven, diagnostic subsystem to ensure that the software will operate in practice as it does in theory and under test. The experiment uses many peripheral devices controlled by the software described here. These devices include: (1) a solid-state data recorder; (2) a bubble memory storage module; (3) a real-time clock; (4) an RS-232C serial interface; (5) a power control subsystem; (6) a matched filter subsystem to detect activation of the Space Shuttle's auxillary power units five minutes prior to launch; (7) a launch detection subsystem based on vibrational and barometric sensors; (8) analog-to-digital converters; and (9) a heater subsystem. The matched filter design is discussed in detail and the results of a computer simulation of the performance of its most critical sub-circuit are presented.

  11. Language experience and consonantal context effects on perceptual assimilation of French vowels by American-English learners of French1

    PubMed Central

    Levy, Erika S.

    2009-01-01

    Recent research has called for an examination of perceptual assimilation patterns in second-language speech learning. This study examined the effects of language learning and consonantal context on perceptual assimilation of Parisian French (PF) front rounded vowels ∕y∕ and ∕œ∕ by American English (AE) learners of French. AE listeners differing in their French language experience (no experience, formal instruction, formal-plus-immersion experience) performed an assimilation task involving PF ∕y, œ, u, o, i, ε, a∕ in bilabial ∕rabVp∕ and alveolar ∕radVt∕ contexts, presented in phrases. PF front rounded vowels were assimilated overwhelmingly to back AE vowels. For PF ∕œ∕, assimilation patterns differed as a function of language experience and consonantal context. However, PF ∕y∕ revealed no experience effect in alveolar context. In bilabial context, listeners with extensive experience assimilated PF ∕y∕ to ∕ju∕ less often than listeners with no or only formal experience, a pattern predicting the poorest ∕u-y∕ discrimination for the most experienced group. An “internal consistency” analysis indicated that responses were most consistent with extensive language experience and in bilabial context. Acoustical analysis revealed that acoustical similarities among PF vowels alone cannot explain context-specific assimilation patterns. Instead it is suggested that native-language allophonic variation influences context-specific perceptual patterns in second-language learning. PMID:19206888

  12. The acoustic properties of bilingual infant-directed speech.

    PubMed

    Danielson, D Kyle; Seidl, Amanda; Onishi, Kristine H; Alamian, Golnoush; Cristia, Alejandrina

    2014-02-01

    Does the acoustic input for bilingual infants equal the conjunction of the input heard by monolinguals of each separate language? The present letter tackles this question, focusing on maternal speech addressed to 11-month-old infants, on the cusp of perceptual attunement. The acoustic characteristics of the point vowels /a,i,u/ were measured in the spontaneous infant-directed speech of French-English bilingual mothers, as well as in the speech of French and English monolingual mothers. Bilingual caregivers produced their two languages with acoustic prosodic separation equal to that of the monolinguals, while also conveying distinct spectral characteristics of the point vowels in their two languages.

  13. Where Do Illusory Vowels Come from?

    ERIC Educational Resources Information Center

    Dupoux, Emmanuel; Parlato, Erika; Frota, Sonia; Hirose, Yuki; Peperkamp, Sharon

    2011-01-01

    Listeners of various languages tend to perceive an illusory vowel inside consonant clusters that are illegal in their native language. Here, we test whether this phenomenon arises after phoneme categorization or rather interacts with it. We assess the perception of illegal consonant clusters in native speakers of Japanese, Brazilian Portuguese,…

  14. Non-Native Production of Thai: Acoustic Measurements and Accentedness Ratings.

    ERIC Educational Resources Information Center

    Wayland, Ratree

    1997-01-01

    Reports a study of the production of Thai vowels, consonants, and tones by native English speakers using two forms of evaluation: acoustic measurements and auditory evaluation by native Thai-speaking listeners. The study focused on the differences in acoustic parameters between the two groups and the acoustic parameter influencing native…

  15. Developmental Dyslexics Show Deficits in the Processing of Temporal Auditory Information in German Vowel Length Discrimination

    ERIC Educational Resources Information Center

    Groth, Katarina; Lachmann, Thomas; Riecker, Axel; Muthmann, Irene; Steinbrink, Claudia

    2011-01-01

    The present study investigated auditory temporal processing in developmental dyslexia by using a vowel length discrimination task. Both temporal and phonological processing were studied in a single experiment. Seven German vowel pairs differing in vowel height were used. The vowels of each pair differed only with respect to vowel length (e.g., /a/…

  16. Gust Acoustic Response of a Single Airfoil Using the Space-Time CE/SE Method

    NASA Technical Reports Server (NTRS)

    Scott, James (Technical Monitor); Wang, X. Y.; Chang, S. C.; Himansu, A.; Jorgenson, P. C. E.

    2003-01-01

    A 2D parallel Euler code based on the space-time conservation element and solution element (CE/SE) method is validated by solving the benchmark problem I in Category 3 of the Third CAA Workshop. This problem concerns the acoustic field generated by the interaction of a convected harmonic vortical gust with a single airfoil. Three gust frequencies, two gust configurations, and three airfoil geometries are considered. Numerical results at both near and far fields are presented and compared with the analytical solutions, a frequency-domain solver GUST3D solutions, and a time-domain high-order Discontinuous Spectral Element Method (DSEM) solutions. It is shown that the CE/SE solutions agree well with the GUST3D solution for the lowest frequency, while there are discrepancies between CE/SE and GUST3D solutions for higher frequencies. However, the CE/SE solution is in good agreement with the DSEM solution for these higher frequencies. It demonstrates that the CE/SE method can produce accurate results of CAA problems involving complex geometries by using unstructured meshes.

  17. Gust Acoustic Response of a Swept Rectilinear Cascade Using The Space-Time CE/SE Method

    NASA Technical Reports Server (NTRS)

    Wang, X. Y.; Himansu, A.; Jorgenson, P. C.; Chang, S. C.

    2001-01-01

    The benchmark problem 3 in Category 3 of the third Computational Aero-Acoustics (CAA) Workshop sponsored by NASA Glenn Research Center is solved using the space-time conservation element and solution element (CE/SE) method. This problem concerns the unsteady response of a rectilinear swept cascade to an incident gust. The acoustic field generated by the interaction of the gust with swept at plates in the cascade is computed by solving the 3D nonlinear Euler equations using the space-time CE/SE method. A parallel version of the 3D CE/SE Euler solver is employed to obtain numerical solutions for several sweep angles. Numerical solutions are presented and compared with the analytical solutions.

  18. Neural Representation of Concurrent Vowels in Macaque Primary Auditory Cortex.

    PubMed

    Fishman, Yonatan I; Micheyl, Christophe; Steinschneider, Mitchell

    2016-01-01

    Successful speech perception in real-world environments requires that the auditory system segregate competing voices that overlap in frequency and time into separate streams. Vowels are major constituents of speech and are comprised of frequencies (harmonics) that are integer multiples of a common fundamental frequency (F0). The pitch and identity of a vowel are determined by its F0 and spectral envelope (formant structure), respectively. When two spectrally overlapping vowels differing in F0 are presented concurrently, they can be readily perceived as two separate "auditory objects" with pitches at their respective F0s. A difference in pitch between two simultaneous vowels provides a powerful cue for their segregation, which in turn, facilitates their individual identification. The neural mechanisms underlying the segregation of concurrent vowels based on pitch differences are poorly understood. Here, we examine neural population responses in macaque primary auditory cortex (A1) to single and double concurrent vowels (/a/ and /i/) that differ in F0 such that they are heard as two separate auditory objects with distinct pitches. We find that neural population responses in A1 can resolve, via a rate-place code, lower harmonics of both single and double concurrent vowels. Furthermore, we show that the formant structures, and hence the identities, of single vowels can be reliably recovered from the neural representation of double concurrent vowels. We conclude that A1 contains sufficient spectral information to enable concurrent vowel segregation and identification by downstream cortical areas. PMID:27294198

  19. The role of consonant/vowel organization in perceptual discrimination.

    PubMed

    Chetail, Fabienne; Drabs, Virginie; Content, Alain

    2014-07-01

    According to a recent hypothesis, the CV pattern (i.e., the arrangement of consonant and vowel letters) constrains the mental representation of letter strings, with each vowel or vowel cluster being the core of a unit. Six experiments with the same/different task were conducted to test whether this structure is extracted prelexically. In the mismatching trials, the targets were pseudowords built by the transposition of 2 adjacent letters from base words. In one condition, the pseudowords had the same number of vowel clusters as the base word, whereas in another condition, the transposition modified the number of vowel clusters (e.g., poirver: 2 vowel clusters vs. povirer: 3 vowel clusters, from POIVRER: 2 vowel clusters). In Experiment 1, pseudowords with a different number of vowel clusters were more quickly processed than pseudowords preserving the CV structure of their base word. Experiment 2 further showed that this effect was not due to changes in syllabic structure. In Experiment 3, the pattern of results was also replicated when the category (consonant or vowel) of the transposed letters was strictly equated between conditions. Experiments 4 and 5 confirmed that the effects were not attributable to lexical processing, to differences in letter identity, or to the position of transpositions. The results suggest that the orthographic representation of letter strings is influenced by the CV pattern at an early, prelexical processing stage.

  20. Vowel categorization during word recognition in bilingual toddlers.

    PubMed

    Ramon-Casas, Marta; Swingley, Daniel; Sebastián-Gallés, Núria; Bosch, Laura

    2009-08-01

    Toddlers' and preschoolers' knowledge of the phonological forms of words was tested in Spanish-learning, Catalan-learning, and bilingual children. These populations are of particular interest because of differences in the Spanish and Catalan vowel systems: Catalan has two vowels in a phonetic region where Spanish has only one. The proximity of the Spanish vowel to the Catalan ones might pose special learning problems. Children were shown picture pairs; the target picture's name was spoken correctly, or a vowel in the target word was altered. Altered vowels either contrasted with the usual vowel in Spanish and Catalan, or only in Catalan. Children's looking to the target picture was used as a measure of word recognition. Monolinguals' word recognition was hindered by within-language, but not non-native, vowel changes. Surprisingly, bilingual toddlers did not show sensitivity to changes in vowels contrastive only in Catalan. Among preschoolers, Catalan-dominant bilinguals but not Spanish-dominant bilinguals revealed mispronunciation sensitivity for the Catalan-only contrast. These studies reveal monolingual children's robust knowledge of native-language vowel categories in words, and show that bilingual children whose two languages contain phonetically overlapping vowel categories may not treat those categories as separate in language comprehension. PMID:19338984

  1. Vowel Categorization during Word Recognition in Bilingual Toddlers

    PubMed Central

    Ramon-Casas, Marta; Swingley, Daniel; Sebastián-Gallés, Núria; Bosch, Laura

    2009-01-01

    Toddlers’ and preschoolers’ knowledge of the phonological forms of words was tested in Spanish-learning, Catalan-learning, and bilingual children. These populations are of particular interest because of differences in the Spanish and Catalan vowel systems: Catalan has two vowels in a phonetic region where Spanish has only one. The proximity of the Spanish vowel to the Catalan ones might pose special learning problems. Children were shown picture pairs; the target picture’s name was spoken correctly, or a vowel in the target word was altered. Altered vowels either contrasted with the usual vowel in Spanish and Catalan, or only in Catalan. Children’s looking to the target picture was used as a measure of word recognition. Monolinguals’ word recognition was hindered by within-language, but not non-native, vowel changes. Surprisingly, bilingual toddlers did not show sensitivity to changes in vowels contrastive only in Catalan. Among preschoolers, Catalan-dominant bilinguals but not Spanish-dominant bilinguals revealed mispronunciation sensitivity for the Catalan-only contrast. These studies reveal monolingual children’s robust knowledge of native-language vowel categories in words, and show that bilingual children whose two languages contain phonetically overlapping vowel categories may not treat those categories as separate in language comprehension. PMID:19338984

  2. Perception of vowels by learners of Spanish and English

    NASA Astrophysics Data System (ADS)

    Garcia-Bayonas, Mariche

    2005-04-01

    This study investigates the perception of English vowels /i I/, /u U/, and /e EI/ and Spanish /i u e/ by native-speakers (NS) and learners (L) and compares these two sets of vowels cross-linguistically. Research on the acquisition of vowels indicates that learners can improve their perception with exposure to the second language [Bohn and Flege (1990)]. Johnson, Flemming, and Wright (1993) investigated the hyperspace effect and how listeners tended to choose extreme vowel qualities in a method of adjustment (MOA) task. The theoretical framework of this study is Fleges (1995) Speech Learning Model. The research question is: Are vowels selected differently by NS and L using synthesized data? Spanish learners (n=54) and English learners (n=17) completed MOA tasks in which they were exposed to 330 synthetically produced vowels to analyze spectral differences in the acquisition of both sound systems, and how the learners vowel system may vary from that of the NS. In the MOA tasks they were asked to select which synthesized vowel sounds resembled the most the ones whose spelling was presented to them. The results include an overview of the vowel formant analysis performed, and which vowels are the most challenging ones to learners.

  3. Evaluation of Space Transportation System (STS) OV-102 orbiter payload bay acoustic environment

    NASA Technical Reports Server (NTRS)

    On, F. J.

    1982-01-01

    This report presents the results of a statistical analysis and evaluation performed by the Mechanical Engineering Branch (Code 731) on the payload bay internal acoustic data measured on the STS-1 through STS-4 flights. The results are used as a basis for developing the required baseline acoustic environment specification to be used in establishing design and test criteria for STS payloads and components.

  4. Existence domains of slow and fast ion-acoustic solitons in two-ion space plasmas

    SciTech Connect

    Maharaj, S. K.; Bharuthram, R.; Singh, S. V. Lakhina, G. S.

    2015-03-15

    A study of large amplitude ion-acoustic solitons is conducted for a model composed of cool and hot ions and cool and hot electrons. Using the Sagdeev pseudo-potential formalism, the scope of earlier studies is extended to consider why upper Mach number limitations arise for slow and fast ion-acoustic solitons. Treating all plasma constituents as adiabatic fluids, slow ion-acoustic solitons are limited in the order of increasing cool ion concentrations by the number densities of the cool, and then the hot ions becoming complex valued, followed by positive and then negative potential double layer regions. Only positive potentials are found for fast ion-acoustic solitons which are limited only by the hot ion number density having to remain real valued. The effect of neglecting as opposed to including inertial effects of the hot electrons is found to induce only minor quantitative changes in the existence regions of slow and fast ion-acoustic solitons.

  5. Arabic Phonology: An Acoustical and Physiological Investigation.

    ERIC Educational Resources Information Center

    Al-Ani, Salman H.

    This book presents an acoustical and physiological Investigation of contemporary standard Arabic as spoken in Iraq. Spectrograms and X-ray sound films are used to perform the analysis for the study. With this equipment, the author considers the vowels, consonants, pharyngealized consonants, pharyngeals and glottals, duration, gemination, and…

  6. Vibration, acoustic, and shock design and test criteria for components on the Solid Rocket Boosters (SRB), Lightweight External Tank (LWT), and Space Shuttle Main Engines (SSME)

    NASA Technical Reports Server (NTRS)

    1984-01-01

    The vibration, acoustics, and shock design and test criteria for components and subassemblies on the space shuttle solid rocket booster (SRB), lightweight tank (LWT), and main engines (SSME) are presented. Specifications for transportation, handling, and acceptance testing are also provided.

  7. Simulating and understanding the effects of velar coupling area on nasalized vowel spectra

    NASA Astrophysics Data System (ADS)

    Pruthi, Tarun; Espy-Wilson, Carol Y.

    2005-09-01

    MRI-based area functions for the nasal cavity of one speaker were combined with the area functions for the vowels /iy/ and /aa/ to study nasalized vowels. The oral cavity was compensated for the falling velum by decreasing the oral cavity area by an amount equal to the increase in the nasal cavity area. Susceptance plots were used along with the simulated transfer functions to understand the effects of velar coupling on nasalized vowel spectra. Susceptance plots of -(Bp+Bo) and Bn suggested significant deviation from the rules suggested by O. Fujimura and J. Lindqvist [J. Acoust. Soc. Am. 49(2), 541-558 (1971)]. In particular, the plots showed that: (1) the frequency of zero crossings of the susceptance plots changes with a change in the coupling area, and (2) formant frequencies need not shift monotonically upward with an increase in coupling area. Further, as a consequence of (1), and the fact that an increase in the coupling area results in a shift of Bn to the right and -(Bp+Bo) to the left, it is postulated that zero crossings of the two plots can cross each other. [MRI data from Brad Story. Work supported by NSF Grant No. BCS0236707.

  8. Prosodic effects on glide-vowel sequences in three Romance languages

    NASA Astrophysics Data System (ADS)

    Chitoran, Ioana

    2001-05-01

    Glide-vowel sequences occur in many Romance languages. In some they can vary in production, ranging from diphthongal pronunciation [ja,je] to hiatus [ia,ie]. According to native speakers' impressionistic perceptions, Spanish and Romanian both exhibit this variation, but to different degrees. Spanish favors glide-vowel sequences, while Romanian favors hiatus, occasionally resulting in different pronunciations of the same items: Spanish (b[j]ela, ind[j]ana), Romanian (b[i]ela, ind[i]ana). The third language, French, has glide-vowel sequences consistently (b[j]elle). This study tests the effect of position in the word on the acoustic duration of the sequences. Shorter duration indicates diphthong production [jV], while longer duration, hiatus [iV]. Eleven speakers (4 Spanish, 4 Romanian, 3 French), were recorded. Spanish and Romanian showed a word position effect. Word-initial sequences were significantly longer than word-medial ones (p<0.001), consistent with native speakers more frequent description of hiatus word-initially than medially. The effect was not found in French (p>0.05). In the Spanish and Romanian sentences, V in the sequence bears pitch accent, but not in French. It is therefore possible that duration is sensitive not to the presence/absence of the word boundary, but to its position relative to pitch accent. The results suggest that the word position effect is crucially enhanced by pitch accent on V.

  9. New design of the pulsed electro-acoustic upper electrode for space charge measurements during electronic irradiation

    NASA Astrophysics Data System (ADS)

    Riffaud, J.; Griseri, V.; Berquez, L.

    2016-07-01

    The behaviour of space charges injected in irradiated dielectrics has been studied for many years for space industry applications. In our case, the pulsed electro-acoustic method is chosen in order to determine the spatial distribution of injected electrons. The feasibility of a ring-shaped electrode which will allow the measurements during irradiation is presented. In this paper, a computer simulation is made in order to determine the parameters to design the electrode and find its position above the sample. The obtained experimental results on polyethylene naphthalate samples realized during electronic irradiation and through relaxation under vacuum will be presented and discussed.

  10. New design of the pulsed electro-acoustic upper electrode for space charge measurements during electronic irradiation.

    PubMed

    Riffaud, J; Griseri, V; Berquez, L

    2016-07-01

    The behaviour of space charges injected in irradiated dielectrics has been studied for many years for space industry applications. In our case, the pulsed electro-acoustic method is chosen in order to determine the spatial distribution of injected electrons. The feasibility of a ring-shaped electrode which will allow the measurements during irradiation is presented. In this paper, a computer simulation is made in order to determine the parameters to design the electrode and find its position above the sample. The obtained experimental results on polyethylene naphthalate samples realized during electronic irradiation and through relaxation under vacuum will be presented and discussed.

  11. Volume I. Percussion Sextet. (original Composition). Volume II. The Simulation of Acoustical Space by Means of Physical Modeling.

    NASA Astrophysics Data System (ADS)

    Manzara, Leonard Charles

    1990-01-01

    The dissertation is in two parts:. 1. Percussion Sextet. The Percussion Sextet is a one movement musical composition with a length of approximately fifteen minutes. It is for six instrumentalists, each on a number of percussion instruments. The overriding formal problem was to construct a coherent and compelling structure which fuses a diversity of musical materials and textures into a dramatic whole. Particularly important is the synthesis of opposing tendencies contained in stochastic and deterministic processes: global textures versus motivic detail, and randomness versus total control. Several compositional techniques are employed in the composition. These methods of composition will be aided, in part, by the use of artificial intelligence techniques programmed on a computer. Finally, the percussion ensemble is the ideal medium to realize the above processes since it encompasses a wide range of both pitched and unpitched timbres, and since a great variety of textures and densities can be created with a certain economy of means. 2. The simulation of acoustical space by means of physical modeling. This is a written report describing the research and development of a computer program which simulates the characteristics of acoustical space in two dimensions. With the computer program the user can simulate most conventional acoustical spaces, as well as those physically impossible to realize in the real world. The program simulates acoustical space by means of geometric modeling. This involves defining wall equations, phantom source points and wall diffusions, and then processing input files containing digital signals through the program, producing output files ready for digital to analog conversion. The user of the program is able to define wall locations and wall reflectivity and roughness characteristics, all of which can be changed over time. Sound source locations are also definable within the acoustical space and these locations can be changed independently at

  12. Formant discrimination in noise for isolated vowels

    NASA Astrophysics Data System (ADS)

    Liu, Chang; Kewley-Port, Diane

    2004-11-01

    Formant discrimination for isolated vowels presented in noise was investigated for normal-hearing listeners. Discrimination thresholds for F1 and F2, for the seven American English vowels /eye, smcapi, eh, æ, invv, aye, you/, were measured under two types of noise, long-term speech-shaped noise (LTSS) and multitalker babble, and also under quiet listening conditions. Signal-to-noise ratios (SNR) varied from -4 to +4 dB in steps of 2 dB. All three factors, formant frequency, signal-to-noise ratio, and noise type, had significant effects on vowel formant discrimination. Significant interactions among the three factors showed that threshold-frequency functions depended on SNR and noise type. The thresholds at the lowest levels of SNR were highly elevated by a factor of about 3 compared to those in quiet. The masking functions (threshold vs SNR) were well described by a negative exponential over F1 and F2 for both LTSS and babble noise. Speech-shaped noise was a slightly more effective masker than multitalker babble, presumably reflecting small benefits (1.5 dB) due to the temporal variation of the babble. .

  13. Vowelling and semantic priming effects in Arabic.

    PubMed

    Mountaj, Nadia; El Yagoubi, Radouane; Himmi, Majid; Lakhdar Ghazal, Faouzi; Besson, Mireille; Boudelaa, Sami

    2015-01-01

    In the present experiment we used a semantic judgment task with Arabic words to determine whether semantic priming effects are found in the Arabic language. Moreover, we took advantage of the specificity of the Arabic orthographic system, which is characterized by a shallow (i.e., vowelled words) and a deep orthography (i.e., unvowelled words), to examine the relationship between orthographic and semantic processing. Results showed faster Reaction Times (RTs) for semantically related than unrelated words with no difference between vowelled and unvowelled words. By contrast, Event Related Potentials (ERPs) revealed larger N1 and N2 components to vowelled words than unvowelled words suggesting that visual-orthographic complexity taxes the early word processing stages. Moreover, semantically unrelated Arabic words elicited larger N400 components than related words thereby demonstrating N400 effects in Arabic. Finally, the Arabic N400 effect was not influenced by orthographic depth. The implications of these results for understanding the processing of orthographic, semantic, and morphological structures in Modern Standard Arabic are discussed. PMID:25528401

  14. Intrinsic fundamental frequency of vowels is moderated by regional dialect.

    PubMed

    Jacewicz, Ewa; Fox, Robert Allen

    2015-10-01

    There has been a long-standing debate whether the intrinsic fundamental frequency (IF0) of vowels is an automatic consequence of articulation or whether it is independently controlled by speakers to perceptually enhance vowel contrasts along the height dimension. This paper provides evidence from regional variation in American English that IF0 difference between high and low vowels is, in part, controlled and varies across dialects. The sources of this F0 control are socio-cultural and cannot be attributed to differences in the vowel inventory size. The socially motivated enhancement was found only in prosodically prominent contexts. PMID:26520352

  15. Intrinsic fundamental frequency of vowels is moderated by regional dialect.

    PubMed

    Jacewicz, Ewa; Fox, Robert Allen

    2015-10-01

    There has been a long-standing debate whether the intrinsic fundamental frequency (IF0) of vowels is an automatic consequence of articulation or whether it is independently controlled by speakers to perceptually enhance vowel contrasts along the height dimension. This paper provides evidence from regional variation in American English that IF0 difference between high and low vowels is, in part, controlled and varies across dialects. The sources of this F0 control are socio-cultural and cannot be attributed to differences in the vowel inventory size. The socially motivated enhancement was found only in prosodically prominent contexts.

  16. A dedicated neural mechanism for vowel selection: a case of relative vowel deficit sparing the number lexicon.

    PubMed

    Semenza, Carlo; Bencini, Giulia M L; Bertella, Laura; Mori, Ileana; Pignatti, Riccardo; Ceriani, Francesca; Cherrick, Danielle; Caldognetto, Emanuela Magno

    2007-01-28

    We report the case of an Italian speaker (GBC), with classical Wernicke's aphasia following a vascular lesion in the posterior middle temporal region. GBC exhibited a selective deficit in spoken language production affecting vowels more than consonants. In reading from a newspaper, GBC substituted vowels for other vowels from the Italian inventory at a rate of 7/1 compared to consonants. No effects of frequency or grammatical class were found. Vowel errors could also not be accounted for by morphological or known phonological processes. Production of number words, in contrast, was free from phonological errors. While GBC has intact representations of Italian vowels and consonants, his data argue for a separate selection mechanism for vowels that is dissociable from that used for consonants. This case provides neuropsychological evidence for models of word production that distinguish between the abstract phonological properties of a word (e.g., sequencing of phonemic slots, or "CV skeleton") and a separate representation for the specific sounds (melody).

  17. Characterization of Pump-Induced Acoustics in Space Launch System Main Propulsion System Liquid Hydrogen Feedline Using Airflow Test Data

    NASA Technical Reports Server (NTRS)

    Eberhart, C. J.; Snellgrove, L. M.; Zoladz, T. F.

    2015-01-01

    High intensity acoustic edgetones located upstream of the RS-25 Low Pressure Fuel Turbo Pump (LPFTP) were previously observed during Space Launch System (STS) airflow testing of a model Main Propulsion System (MPS) liquid hydrogen (LH2) feedline mated to a modified LPFTP. MPS hardware has been adapted to mitigate the problematic edgetones as part of the Space Launch System (SLS) program. A follow-on airflow test campaign has subjected the adapted hardware to tests mimicking STS-era airflow conditions, and this manuscript describes acoustic environment identification and characterization born from the latest test results. Fluid dynamics responsible for driving discrete excitations were well reproduced using legacy hardware. The modified design was found insensitive to high intensity edgetone-like discretes over the bandwidth of interest to SLS MPS unsteady environments. Rather, the natural acoustics of the test article were observed to respond in a narrowband-random/mixed discrete manner to broadband noise thought generated by the flow field. The intensity of these responses were several orders of magnitude reduced from those driven by edgetones.

  18. Arbitrary amplitude fast electron-acoustic solitons in three-electron component space plasmas

    NASA Astrophysics Data System (ADS)

    Mbuli, L. N.; Maharaj, S. K.; Bharuthram, R.; Singh, S. V.; Lakhina, G. S.

    2016-06-01

    We examine the characteristics of fast electron-acoustic solitons in a four-component unmagnetised plasma model consisting of cool, warm, and hot electrons, and cool ions. We retain the inertia and pressure for all the plasma species by assuming adiabatic fluid behaviour for all the species. By using the Sagdeev pseudo-potential technique, the allowable Mach number ranges for fast electron-acoustic solitary waves are explored and discussed. It is found that the cool and warm electron number densities determine the polarity switch of the fast electron-acoustic solitons which are limited by either the occurrence of fast electron-acoustic double layers or warm and hot electron number density becoming unreal. For the first time in the study of solitons, we report on the coexistence of fast electron-acoustic solitons, in addition to the regular fast electron-acoustic solitons and double layers in our multi-species plasma model. Our results are applied to the generation of broadband electrostatic noise in the dayside auroral region.

  19. The Role of Secondary-Stressed and Unstressed-Unreduced Syllables in Word Recognition: Acoustic and Perceptual Studies with Russian Learners of English

    ERIC Educational Resources Information Center

    Banzina, Elina; Dilley, Laura C.; Hewitt, Lynne E.

    2016-01-01

    The importance of secondary-stressed (SS) and unstressed-unreduced (UU) syllable accuracy for spoken word recognition in English is as yet unclear. An acoustic study first investigated Russian learners' of English production of SS and UU syllables. Significant vowel quality and duration reductions in Russian-spoken SS and UU vowels were found,…

  20. Spectral analysis of sung vowels. I. Variation due to differences between vowels, singers, and modes of singing.

    PubMed

    Bloothooft, G; Plomp, R

    1984-04-01

    Average 1/3-oct filter spectra of vowels, sung by seven professional male and seven professional female singers , were measured. The material consisted of nine different vowels, sung at six fundamental frequencies (F0, ranging from 98 up to 880 Hz). For each vowel the singers were requested to sing in the following nine modes: neutral, light, dark, pressed, free, loud, soft, straight, and extra vibrato . To study origins of spectral variation, the quantity of spectral variance, based on band filter sound levels, was used. For each fundamental frequency and separately for males and females, portions of total spectral variance associated with the main effects and the interactions of the factors "vowels," " singers ," and "modes of singing," were computed. A considerable decrease in total spectral variance was found when F0 rose from 98 to 880 Hz, mostly due to the reduced spectral variance between vowels. Above about F0 = 660 Hz spectral variation was dominated by differences related to singers and modes of singing. Additional analyses revealed that for all F0 values vowel spectra of the tenor and the soprano singers varied more than those of the bass and the alto singers , there was only a slight dependence of spectral differences between vowels on the mode of singing, and the amount of spectral variation in a vowel, sung by different singers with different modes of singing, was vowel dependent. PMID:6725777

  1. Radiometric and photometric design for an Acoustic Containerless Experiment System. [for space processing

    NASA Technical Reports Server (NTRS)

    Glavich, T. A.

    1981-01-01

    The design of an optical system for a high temperature Acoustic Containerless Experiment System is examined. The optical system provides two-axis video, cine and infrared images of an acoustically positioned sample over a temperature range of 20 to 1200 C. Emphasis is placed on the radiometric and photometric characterization of the elements in the optical system and the oven to assist image data determination. Sample visibility due to wall radiance is investigated along with visibility due to strobe radiance. The optical system is designed for operation in Spacelab, and is used for a variety of materials processing experiments.

  2. A k-space method for acoustic propagation using coupled first-order equations in three dimensions

    PubMed Central

    Tillett, Jason C.; Daoud, Mohammad I.; Lacefield, James C.; Waag, Robert C.

    2009-01-01

    A previously described two-dimensional k-space method for large-scale calculation of acoustic wave propagation in tissues is extended to three dimensions. The three-dimensional method contains all of the two-dimensional method features that allow accurate and stable calculation of propagation. These features are spectral calculation of spatial derivatives, temporal correction that produces exact propagation in a homogeneous medium, staggered spatial and temporal grids, and a perfectly matched boundary layer. Spectral evaluation of spatial derivatives is accomplished using a fast Fourier transform in three dimensions. This computational bottleneck requires all-to-all communication; execution time in a parallel implementation is therefore sensitive to node interconnect latency and bandwidth. Accuracy of the three-dimensional method is evaluated through comparisons with exact solutions for media having spherical inhomogeneities. Large-scale calculations in three dimensions were performed by distributing the nearly 50 variables per voxel that are used to implement the method over a cluster of computers. Two computer clusters used to evaluate method accuracy are compared. Comparisons of k-space calculations with exact methods including absorption highlight the need to model accurately the medium dispersion relationships, especially in large-scale media. Accurately modeled media allow the k-space method to calculate acoustic propagation in tissues over hundreds of wavelengths. PMID:19739736

  3. Vowel Categorization during Word Recognition in Bilingual Toddlers

    ERIC Educational Resources Information Center

    Ramon-Casas, Marta; Swingley, Daniel; Sebastian-Galles, Nuria; Bosch, Laura

    2009-01-01

    Toddlers' and preschoolers' knowledge of the phonological forms of words was tested in Spanish-learning, Catalan-learning, and bilingual children. These populations are of particular interest because of differences in the Spanish and Catalan vowel systems: Catalan has two vowels in a phonetic region where Spanish has only one. The proximity of the…

  4. Mommy, Speak Clearly: Induced Hearing Loss Shapes Vowel Hyperarticulation

    ERIC Educational Resources Information Center

    Lam, Christa; Kitamura, Christine

    2012-01-01

    Talkers hyperarticulate vowels when communicating with listeners that require increased speech intelligibility. Vowel hyperarticulation is said to be motivated by knowledge of the listener's linguistic needs because it typically occurs in speech to infants, foreigners and hearing-impaired listeners, but not to non-verbal pets. However, the degree…

  5. Native Italian speakers' perception and production of English vowels.

    PubMed

    Flege, J E; MacKay, I R; Meador, D

    1999-11-01

    This study examined the production and perception of English vowels by highly experienced native Italian speakers of English. The subjects were selected on the basis of the age at which they arrived in Canada and began to learn English, and how much they continued to use Italian. Vowel production accuracy was assessed through an intelligibility test in which native English-speaking listeners attempted to identify vowels spoken by the native Italian subjects. Vowel perception was assessed using a categorial discrimination test. The later in life the native Italian subjects began to learn English, the less accurately they produced and perceived English vowels. Neither of two groups of early Italian/English bilinguals differed significantly from native speakers of English either for production or perception. This finding is consistent with the hypothesis of the speech learning model [Flege, in Speech Perception and Linguistic Experience: Theoretical and Methodological Issues (York, Timonium, MD, 1995)] that early bilinguals establish new categories for vowels found in the second language (L2). The significant correlation observed to exist between the measures of L2 vowel production and perception is consistent with another hypothesis of the speech learning model, viz., that the accuracy with which L2 vowels are produced is limited by how accurately they are perceived. PMID:10573909

  6. Vowel Perception and Production in Adolescents with Reading Disabilities.

    ERIC Educational Resources Information Center

    Bertucci, Carol; Hook, Pamela; Haynes, Charles; Macaruso, Paul; Bickley, Corine

    2003-01-01

    Perception and production of vowels in the words "pit,""pet," and "pat" were investigated with 19 adolescents with reading disabilities. Students with reading disabilities perceived and produced less well-defined vowel categories than a control group. Results suggest that speech processing difficulties of students with reading disabilities include…

  7. Congruent and Incongruent Semantic Context Influence Vowel Recognition

    ERIC Educational Resources Information Center

    Wotton, J. M.; Elvebak, R. L.; Moua, L. C.; Heggem, N. M.; Nelson, C. A.; Kirk, K. M.

    2011-01-01

    The influence of sentence context on the recognition of naturally spoken vowels degraded by reverberation and Gaussian noise was investigated. Target words were paired to have similar consonant sounds but different vowels (e.g., map/mop) and were embedded early in sentences which provided three types of semantic context. Fifty-eight…

  8. Variation in vowel duration among southern African American English speakers

    PubMed Central

    Holt, Yolanda Feimster; Jacewicz, Ewa; Fox, Robert Allen

    2015-01-01

    Purpose Atypical duration of speech segments can signal a speech disorder. This study examined variation in vowel duration in African American English (AAE) relative to White American English (WAE) speakers living in the same dialect region in the South in order to characterize the nature of systematic variation between the two groups. The goal was to establish whether segmental durations in minority populations differ from the well-established patterns in mainstream populations. Method Participants were 32 AAE and 32 WAE speakers differing in age who, in their childhood, attended either segregated (older speakers) or integrated (younger speakers) public schools. Speech materials consisted of 14 vowels produced in hVd-frame. Results AAE vowels were significantly longer than WAE vowels. Vowel duration did not differ as a function of age. The temporal tense-lax contrast was minimized for AAE relative to WAE. Female vowels were significantly longer than male vowels for both AAE and WAE. Conclusions African Americans should be expected to produce longer vowels relative to White speakers in a common geographic area. These longer durations are not deviant but represent a typical feature of AAE. This finding has clinical importance in guiding assessments of speech disorders in AAE speakers. PMID:25951511

  9. Bite Block Vowel Production in Apraxia of Speech

    ERIC Educational Resources Information Center

    Jacks, Adam

    2008-01-01

    Purpose: This study explored vowel production and adaptation to articulatory constraints in adults with acquired apraxia of speech (AOS) plus aphasia. Method: Five adults with acquired AOS plus aphasia and 5 healthy control participants produced the vowels [iota], [epsilon], and [ash] in four word-length conditions in unconstrained and bite block…

  10. Discrimination of Phonemic Vowel Length by Japanese Infants

    ERIC Educational Resources Information Center

    Sato, Yutaka; Sogabe, Yuko; Mazuka, Reiko

    2010-01-01

    Japanese has a vowel duration contrast as one component of its language-specific phonemic repertory to distinguish word meanings. It is not clear, however, how a sensitivity to vowel duration can develop in a linguistic context. In the present study, using the visual habituation-dishabituation method, the authors evaluated infants' abilities to…

  11. Dialect effects in speech perception: the role of vowel duration in Parisian French and Swiss French.

    PubMed

    Miller, Joanne L; Mondini, Michèle; Grosjean, François; Dommergues, Jean-Yves

    2011-12-01

    The current experiments examined how native Parisian French and native Swiss French listeners use vowel duration in perceiving the /[character: see text]/-/o/ contrast. In both Parisian and Swiss French /ol is longer than /[character: see text]/, but the difference is relatively large in Swiss French and quite small in Parisian French. In Experiment I we found a parallel effect in perception. For native listeners of both dialects, the perceived best exemplars of /o/ were longer than those of /[character: see text]/. However, there was a substantial difference in best-exemplar duration for /[character: see text]/ and /o/ for Swiss French listeners, but only a small difference in best-exemplar duration for Parisian French listeners. In Experiment 2 we found that this precise pattern depended not only on the native dialect of the listeners, but also on whether the stimuli being judged had the detailed acoustic characteristics of the native dialect. These findings indicate that listeners use fine-grained information in the speech signal in a dialect-specific manner when mapping the acoustic signal onto vowel categories of their language.

  12. Correlations of decision weights and cognitive function for the masked discrimination of vowels by young and old adults

    PubMed Central

    Lutfi, Robert A.

    2014-01-01

    Older adults are often reported in the literature to have greater difficulty than younger adults understanding speech in noise [Helfer and Wilber (1988). J. Acoust. Soc. Am, 859–893]. The poorer performance of older adults has been attributed to a general deterioration of cognitive processing, deterioration of cochlear anatomy, and/or greater difficulty segregating speech from noise. The current work used perturbation analysis [Berg (1990). J. Acoust. Soc. Am., 149–158] to provide a more specific assessment of the effect of cognitive factors on speech perception in noise. Sixteen older (age 56–79 years) and seventeen younger (age 19–30 years) adults discriminated a target vowel masked by randomly selected masker vowels immediately preceding and following the target. Relative decision weights on target and maskers resulting from the analysis revealed large individual differences across participants despite similar performance scores in many cases. On the most difficult vowel discriminations, the older adult decision weights were significantly correlated with inhibitory control (Color Word Interference test) and pure-tone threshold averages (PTA). Young adult decision weights were not correlated with any measures of peripheral (PTA) or central function (inhibition or working memory). PMID:25256580

  13. Correlations of decision weights and cognitive function for the masked discrimination of vowels by young and old adults.

    PubMed

    Gilbertson, Lynn; Lutfi, Robert A

    2014-11-01

    Older adults are often reported in the literature to have greater difficulty than younger adults understanding speech in noise [Helfer and Wilber (1988). J. Acoust. Soc. Am, 859-893]. The poorer performance of older adults has been attributed to a general deterioration of cognitive processing, deterioration of cochlear anatomy, and/or greater difficulty segregating speech from noise. The current work used perturbation analysis [Berg (1990). J. Acoust. Soc. Am., 149-158] to provide a more specific assessment of the effect of cognitive factors on speech perception in noise. Sixteen older (age 56-79 years) and seventeen younger (age 19-30 years) adults discriminated a target vowel masked by randomly selected masker vowels immediately preceding and following the target. Relative decision weights on target and maskers resulting from the analysis revealed large individual differences across participants despite similar performance scores in many cases. On the most difficult vowel discriminations, the older adult decision weights were significantly correlated with inhibitory control (Color Word Interference test) and pure-tone threshold averages (PTA). Young adult decision weights were not correlated with any measures of peripheral (PTA) or central function (inhibition or working memory).

  14. The multidimensional nature of hyperspeech: evidence from Japanese vowel devoicing.

    PubMed

    Martin, Andrew; Utsugi, Akira; Mazuka, Reiko

    2014-08-01

    We investigate the hypothesis that infant-directed speech is a form of hyperspeech, optimized for intelligibility, by focusing on vowel devoicing in Japanese. Using a corpus of infant-directed and adult-directed Japanese, we show that speakers implement high vowel devoicing less often when speaking to infants than when speaking to adults, consistent with the hyperspeech hypothesis. The same speakers, however, increase vowel devoicing in careful, read speech, a speech style which might be expected to pattern similarly to infant-directed speech. We argue that both infant-directed and read speech can be considered listener-oriented speech styles-each is optimized for the specific needs of its intended listener. We further show that in non-high vowels, this trend is reversed: speakers devoice more often in infant-directed speech and less often in read speech, suggesting that devoicing in the two types of vowels is driven by separate mechanisms in Japanese. PMID:24813573

  15. The multidimensional nature of hyperspeech: evidence from Japanese vowel devoicing.

    PubMed

    Martin, Andrew; Utsugi, Akira; Mazuka, Reiko

    2014-08-01

    We investigate the hypothesis that infant-directed speech is a form of hyperspeech, optimized for intelligibility, by focusing on vowel devoicing in Japanese. Using a corpus of infant-directed and adult-directed Japanese, we show that speakers implement high vowel devoicing less often when speaking to infants than when speaking to adults, consistent with the hyperspeech hypothesis. The same speakers, however, increase vowel devoicing in careful, read speech, a speech style which might be expected to pattern similarly to infant-directed speech. We argue that both infant-directed and read speech can be considered listener-oriented speech styles-each is optimized for the specific needs of its intended listener. We further show that in non-high vowels, this trend is reversed: speakers devoice more often in infant-directed speech and less often in read speech, suggesting that devoicing in the two types of vowels is driven by separate mechanisms in Japanese.

  16. Arbitrary amplitude slow electron-acoustic solitons in three-electron temperature space plasmas

    SciTech Connect

    Mbuli, L. N.; Maharaj, S. K.; Bharuthram, R.; Singh, S. V.; Lakhina, G. S.

    2015-06-15

    We examine the characteristics of large amplitude slow electron-acoustic solitons supported in a four-component unmagnetised plasma composed of cool, warm, hot electrons, and cool ions. The inertia and pressure for all the species in this plasma system are retained by assuming that they are adiabatic fluids. Our findings reveal that both positive and negative potential slow electron-acoustic solitons are supported in the four-component plasma system. The polarity switch of the slow electron-acoustic solitons is determined by the number densities of the cool and warm electrons. Negative potential solitons, which are limited by the cool and warm electron number densities becoming unreal and the occurrence of negative potential double layers, are found for low values of the cool electron density, while the positive potential solitons occurring for large values of the cool electron density are only limited by positive potential double layers. Both the lower and upper Mach numbers for the slow electron-acoustic solitons are computed and discussed.

  17. The acquisition of Taiwan Mandarin vowels by native American English speakers

    NASA Astrophysics Data System (ADS)

    Lin, Cyun-Jhan

    2005-04-01

    Previous work on the production of English and French phones by native American English speakers indicated that equivalence classification prevent L2 learners from approximating L2 phonetic norms of similar phones and that learning French would not affect English speakers' production of L1 similar phone /u/ (Flege, 1987). In this study, there were five subjects, including 2 advanced native American English learners of Taiwan Mandarin, 2 basic native American English learners of Taiwan Mandarin, and 1 monolingual Taiwan Mandarin speaker. The corpus were 12 English words ``heed, who'd, hod; leak, Luke, lock; beat, suit, bot; peat, suit, pot,'' and 12 Mandarin words [i,u, a; li, lu, la; pi, pu, pa; phi, phu, pha]. Both advanced and basic learners' production of English and Mandarin words and monolingual Taiwan Mandarin speaker's production of Mandarin words were directly recorded onto a PC. Vowel formants were taken from spectrograms generated by Praat. Preliminary results showed the vowel space of advanced learners between Taiwan Mandarin [i] and [u] was larger than that of basic learners, and closer to the Taiwan Mandarin norms. Besides, the vowel space between English [i] and [u] by basic learners was dramatically smaller than that of American English norms.

  18. Perception of Vowel Length by Japanese- and English-Learning Infants

    ERIC Educational Resources Information Center

    Mugitani, Ryoko; Pons, Ferran; Fais, Laurel; Dietrich, Christiane; Werker, Janet F.; Amano, Shigeaki

    2009-01-01

    This study investigated vowel length discrimination in infants from 2 language backgrounds, Japanese and English, in which vowel length is either phonemic or nonphonemic. Experiment 1 revealed that English 18-month-olds discriminate short and long vowels although vowel length is not phonemically contrastive in English. Experiments 2 and 3 revealed…

  19. The Spelling of Vowels Is Influenced by Australian and British English Dialect Differences

    ERIC Educational Resources Information Center

    Kemp, Nenagh

    2009-01-01

    Two experiments examined the influence of dialect on the spelling of vowel sounds. British and Australian children (6 to 8 years) and university students wrote words whose unstressed vowel sound is spelled i or e and pronounced /I/ or /schwa/. Participants often (mis)spelled these vowel sounds as they pronounced them. When vowels were pronounced…

  20. A Comparative Study of English and Spanish Vowel Systems: Theoretical and Practical Implications for Teaching Pronunciation.

    ERIC Educational Resources Information Center

    Odisho, Edward Y.

    A study examines two major types of vowel systems in languages, centripetal and centrifugal. English is associated with the centripetal system, in which vowel quality and quantity (rhythm) are heavily influenced by stress. In this system, vowels have a strong tendency to move toward the center of the vowel area. Spanish is associated with the…

  1. We're Not in Kansas Anymore: The TOTO Strategy for Decoding Vowel Pairs

    ERIC Educational Resources Information Center

    Meese, Ruth Lyn

    2016-01-01

    Vowel teams such as vowel digraphs present a challenge to struggling readers. Some researchers assert that phonics generalizations such as the "two vowels go walking and the first one does the talking" rule do not hold often enough to be reliable for children. Others suggest that some vowel teams are highly regular and that children can…

  2. Acoustic-phonetic correlates of talker intelligibility for adults and children

    NASA Astrophysics Data System (ADS)

    Hazan, Valerie; Markham, Duncan

    2004-11-01

    This study investigated acoustic-phonetic correlates of intelligibility for adult and child talkers, and whether the relative intelligibility of different talkers was dependent on listener characteristics. In experiment 1, word intelligibility was measured for 45 talkers (18 women, 15 men, 6 boys, 6 girls) from a homogeneous accent group. The material consisted of 124 words familiar to 7-year-olds that adequately covered all frequent consonant confusions; stimuli were presented to 135 adult and child listeners in low-level background noise. Seven-to-eight-year-old listeners made significantly more errors than 12-year-olds or adults, but the relative intelligibility of individual talkers was highly consistent across groups. In experiment 2, listener ratings on a number of voice dimensions were obtained for the adults talkers identified in experiment 1 as having the highest and lowest intelligibility. Intelligibility was significantly correlated with subjective dimensions reflecting articulation, voice dynamics, and general quality. Finally, in experiment 3, measures of fundamental frequency, long-term average spectrum, word duration, consonant-vowel intensity ratio, and vowel space size were obtained for all talkers. Overall, word intelligibility was significantly correlated with the total energy in the 1- to 3-kHz region and word duration; these measures predicted 61% of the variability in intelligibility. The fact that the relative intelligibility of individual talkers was remarkably consistent across listener age groups suggests that the acoustic-phonetic characteristics of a talker's utterance are the primary factor in determining talker intelligibility. Although some acoustic-phonetic correlates of intelligibility were identified, variability in the profiles of the ``best'' talkers suggests that high intelligibility can be achieved through a combination of different acoustic-phonetic characteristics. .

  3. Optimal Transducer Arrangement for Temperature Distribution Measurement in Arbitrary Convex-Shaped Space by Acoustic Computerized Tomography

    NASA Astrophysics Data System (ADS)

    Ayumu Minamide,; Naoto Wakatsuki,; Koichi Mizutani,

    2010-07-01

    We study the optimal transducer arrangement for measuring temperature distribution in an arbitrary convex-shaped space by acoustic computerized tomography (A-CT) using flexibly arranged transducers. The transducer arrangement is optimized by the combined use of two methods, namely, real-coded genetic algorithm (RCGA) and simulated annealing (SA). By RCGA, the optimized arrangement is globally searched and then locally searched by SA. A fitness function of these methods is defined to evaluate the distribution of projection data in θ-r space. By use of this function, we aim to obtain better projection data sets by two-dimensionally interpolating the projection data. By numerical simulations, we confirmed the adequacy of our method and the optimized transducer arrangement.

  4. The acquisition of unrounded vowels in English.

    PubMed

    Otomo, K; Stoel-Gammon, C

    1992-06-01

    This study investigated developmental patterns of acquisition of the unrounded American English vowels /i, I, e, e, ae, a/ by following 6 normally developing children from 22 to 30 months of age. The subjects were examined at approximately 22, 26, and 30 months of age. Results showed that, in general, /i/ and /a/ were mastered early and /I/ and /e/ were least accurate throughout the period of the study. Upon inspection of errors, the following three classes of production errors were identified: (a) intertrial production variability, (b) context-sensitive substitutions, and (c) context-free systematic substitution patterns, or articulatory processes. A decrease in production variability and in the occurrence of articulatory processes with age generally coincided with a gradual improvement in accuracy of production. However, substitutions of lower vowels for /I/ were persistent, and the pattern was observed even at 30 months of age. Individual variation was also evident in the production accuracy, the substitution patterns, and the manner of articulatory improvement. PMID:1608252

  5. Engaging spaces: Intimate electro-acoustic display in alternative performance venues

    NASA Astrophysics Data System (ADS)

    Bahn, Curtis; Moore, Stephan

    2001-05-01

    In past presentations to the ASA, we have described the design and construction of four generations of unique spherical speakers (multichannel, outward-radiating geodesic speaker arrays) and Sensor-Speaker-Arrays, (SenSAs: combinations of various sensor devices with outward-radiating multichannel speaker arrays). This presentation will detail the ways in which arrays of these speakers have been employed in alternative performance venues-providing presence and intimacy in the performance of electro-acoustic chamber music and sound installation, while engaging natural and unique acoustical qualities of various locations. We will present documentation of the use of multichannel sonic diffusion arrays in small clubs, ``black-box'' theaters, planetariums, and art galleries.

  6. Sound in ecclesiastical spaces in Cordoba. Architectural projects incorporating acoustic methodology (El sonido del espacio eclesial en Cordoba. El proyecto arquitectonico como procedimiento acustico)

    NASA Astrophysics Data System (ADS)

    Suarez, Rafael

    2003-11-01

    This thesis is concerned with the acoustic analysis of ecclesiastical spaces, and the subsequent implementation of acoustic design methodology in architectural renovations. One begins with an adequate architectural design of specific elements (shape, materials, and textures), with the intention of elimination of acoustic deficiencies that are common in such spaces. These are those deficiencies that impair good speech intelligibility and good musical audibility. The investigation is limited to churches in the province of Cordoba and to churches built after the reconquest of Spain (1236) and up until the 18th century. Selected churches are those that have undergone architectural renovations to adapt them to new uses or to make them more suitable for liturgical use. The thesis attempts to summarize the acoustic analyses and the acoustical solutions that have been implemented. The results are presented in a manner that should be useful for the adoption of a model for the functional renovation of ecclesiastical spaces. Such would allow those involved in architectural projects to specify the nature of the sound, even though somewhat intangible, within the ecclesiastical space. Thesis advisors: Jaime Navarro and Juan J. Sendra Copies of this thesis written in Spanish may be obtained by contacting the advisor, Jaime Navarro, E.T.S. de Arquitectura de Sevilla, Dpto. de Construcciones Arquitectonicas I, Av. Reina Mercedes, 2, 41012 Sevilla, Spain. E-mail address: jnavarro@us.es

  7. The Acoustic Voice Quality Index: Toward Improved Treatment Outcomes Assessment in Voice Disorders

    ERIC Educational Resources Information Center

    Maryn, Youri; De Bodt, Marc; Roy, Nelson

    2010-01-01

    Voice practitioners require an objective index of dysphonia severity as a means to reliably track treatment outcomes. To ensure ecological validity however, such a measure should survey both sustained vowels and continuous speech. In an earlier study, a multivariate acoustic model referred to as the Acoustic Voice Quality Index (AVQI), consisting…

  8. Phoneme recognition and confusions with multichannel cochlear implants: vowels.

    PubMed

    Välimaa, Taina T; Määttä, Taisto K; Löppönen, Heikki J; Sorri, Martti J

    2002-10-01

    The aim of this study was to investigate how postlingually severely or profoundly hearing-impaired adults relearn to recognize vowels after receiving multichannel cochlear implants. Vowel recognition of 19 Finnish-speaking subjects was studied for a minimum of 6 months and a maximum of 24 months using an open-set nonsense-syllable test in a prospective repeated-measure design. The responses were coded for phoneme errors, and 95% confidence intervals for recognition and confusions were calculated. The average vowel recognition was 68% (95% confidence interval = 66-70%) 6 months after switch-on and 80% (95% confidence interval = 78-82%) 24 months after switch-on. The vowels [ae], [u], [i], [o], and [a] were the easiest to recognize, and the vowels [y], [e], and [ø] were the most difficult. In conclusion, adaptation to electrical hearing using a multichannel cochlear implant was achieved well; but for at least 2 years, given two vowels with either F1 or F2 at roughly the some frequencies, confusions were drawn more towards the closest vowel with the next highest F1 or F2.

  9. Dynamic Spectral Structure Specifies Vowels for Adults and Children

    PubMed Central

    Nittrouer, Susan; Lowenstein, Joanna H.

    2014-01-01

    The dynamic specification account of vowel recognition suggests that formant movement between vowel targets and consonant margins is used by listeners to recognize vowels. This study tested that account by measuring contributions to vowel recognition of dynamic (i.e., time-varying) spectral structure and coarticulatory effects on stationary structure. Adults and children (four-and seven-year-olds) were tested with three kinds of consonant-vowel-consonant syllables: (1) unprocessed; (2) sine waves that preserved both stationary coarticulated and dynamic spectral structure; and (3) vocoded signals that primarily preserved that stationary, but not dynamic structure. Sections of two lengths were removed from syllable middles: (1) half the vocalic portion; and (2) all but the first and last three pitch periods. Adults performed accurately with unprocessed and sine-wave signals, as long as half the syllable remained; their recognition was poorer for vocoded signals, but above chance. Seven-year-olds performed more poorly than adults with both sorts of processed signals, but disproportionately worse with vocoded than sine-wave signals. Most four-year-olds were unable to recognize vowels at all with vocoded signals. Conclusions were that both dynamic and stationary coarticulated structures support vowel recognition for adults, but children attend to dynamic spectral structure more strongly because early phonological organization favors whole words. PMID:25536845

  10. Speaking Clearly for the Blind: Acoustic and Articulatory Correlates of Speaking Conditions in Sighted and Congenitally Blind Speakers.

    PubMed

    Ménard, Lucie; Trudeau-Fisette, Pamela; Côté, Dominique; Turgeon, Christine

    2016-01-01

    Compared to conversational speech, clear speech is produced with longer vowel duration, greater intensity, increased contrasts between vowel categories, and decreased dispersion within vowel categories. Those acoustic correlates are produced by larger movements of the orofacial articulators, including visible (lips) and invisible (tongue) articulators. Thus, clear speech provides the listener with audible and visual cues that are used to increase the overall intelligibility of speech produced by the speaker. It is unclear how those cues are produced by visually impaired speakers who never had access to vision. In this paper, we investigate the acoustic and articulatory correlates of vowels in clear versus conversational speech, and in sighted and congenitally blind speakers. Participants were recorded using electroarticulography while producing multiple repetitions of the ten Quebec French oral vowels in carrier sentences in both speaking conditions. Articulatory variables (lip, jaw, and tongue positions) as well as acoustic variables (contrasts between vowels, within-vowel dispersion, pitch, duration, and intensity) were measured. Lip movements were larger when going from conversational to clear speech in sighted speakers only. On the other hand, tongue movements were affected to a larger extent in blind speakers compared to their sighted peers. These findings confirm that vision plays an important role in the maintenance of speech intelligibility.

  11. Speaking Clearly for the Blind: Acoustic and Articulatory Correlates of Speaking Conditions in Sighted and Congenitally Blind Speakers.

    PubMed

    Ménard, Lucie; Trudeau-Fisette, Pamela; Côté, Dominique; Turgeon, Christine

    2016-01-01

    Compared to conversational speech, clear speech is produced with longer vowel duration, greater intensity, increased contrasts between vowel categories, and decreased dispersion within vowel categories. Those acoustic correlates are produced by larger movements of the orofacial articulators, including visible (lips) and invisible (tongue) articulators. Thus, clear speech provides the listener with audible and visual cues that are used to increase the overall intelligibility of speech produced by the speaker. It is unclear how those cues are produced by visually impaired speakers who never had access to vision. In this paper, we investigate the acoustic and articulatory correlates of vowels in clear versus conversational speech, and in sighted and congenitally blind speakers. Participants were recorded using electroarticulography while producing multiple repetitions of the ten Quebec French oral vowels in carrier sentences in both speaking conditions. Articulatory variables (lip, jaw, and tongue positions) as well as acoustic variables (contrasts between vowels, within-vowel dispersion, pitch, duration, and intensity) were measured. Lip movements were larger when going from conversational to clear speech in sighted speakers only. On the other hand, tongue movements were affected to a larger extent in blind speakers compared to their sighted peers. These findings confirm that vision plays an important role in the maintenance of speech intelligibility. PMID:27643997

  12. Speaking Clearly for the Blind: Acoustic and Articulatory Correlates of Speaking Conditions in Sighted and Congenitally Blind Speakers

    PubMed Central

    Ménard, Lucie; Trudeau-Fisette, Pamela; Côté, Dominique; Turgeon, Christine

    2016-01-01

    Compared to conversational speech, clear speech is produced with longer vowel duration, greater intensity, increased contrasts between vowel categories, and decreased dispersion within vowel categories. Those acoustic correlates are produced by larger movements of the orofacial articulators, including visible (lips) and invisible (tongue) articulators. Thus, clear speech provides the listener with audible and visual cues that are used to increase the overall intelligibility of speech produced by the speaker. It is unclear how those cues are produced by visually impaired speakers who never had access to vision. In this paper, we investigate the acoustic and articulatory correlates of vowels in clear versus conversational speech, and in sighted and congenitally blind speakers. Participants were recorded using electroarticulography while producing multiple repetitions of the ten Quebec French oral vowels in carrier sentences in both speaking conditions. Articulatory variables (lip, jaw, and tongue positions) as well as acoustic variables (contrasts between vowels, within-vowel dispersion, pitch, duration, and intensity) were measured. Lip movements were larger when going from conversational to clear speech in sighted speakers only. On the other hand, tongue movements were affected to a larger extent in blind speakers compared to their sighted peers. These findings confirm that vision plays an important role in the maintenance of speech intelligibility. PMID:27643997

  13. A narrow band pattern-matching model of vowel perception

    NASA Astrophysics Data System (ADS)

    Hillenbrand, James M.; Houde, Robert A.

    2003-02-01

    The purpose of this paper is to propose and evaluate a new model of vowel perception which assumes that vowel identity is recognized by a template-matching process involving the comparison of narrow band input spectra with a set of smoothed spectral-shape templates that are learned through ordinary exposure to speech. In the present simulation of this process, the input spectra are computed over a sufficiently long window to resolve individual harmonics of voiced speech. Prior to template creation and pattern matching, the narrow band spectra are amplitude equalized by a spectrum-level normalization process, and the information-bearing spectral peaks are enhanced by a ``flooring'' procedure that zeroes out spectral values below a threshold function consisting of a center-weighted running average of spectral amplitudes. Templates for each vowel category are created simply by averaging the narrow band spectra of like vowels spoken by a panel of talkers. In the present implementation, separate templates are used for men, women, and children. The pattern matching is implemented with a simple city-block distance measure given by the sum of the channel-by-channel differences between the narrow band input spectrum (level-equalized and floored) and each vowel template. Spectral movement is taken into account by computing the distance measure at several points throughout the course of the vowel. The input spectrum is assigned to the vowel template that results in the smallest difference accumulated over the sequence of spectral slices. The model was evaluated using a large database consisting of 12 vowels in /hVd/ context spoken by 45 men, 48 women, and 46 children. The narrow band model classified vowels in this database with a degree of accuracy (91.4%) approaching that of human listeners.

  14. Vocal Attack Time of Different Pitch Levels and Vowels in Mandarin.

    PubMed

    Zhang, Ruifeng; Baken, R J; Kong, Jiangping

    2015-09-01

    The purpose of this study was to investigate how vocal attack time (VAT) varies when young adults articulate the three vertex vowels in Mandarin Chinese at five linguistically unconstrained pitch levels. Sound pressure and electroglottographic signals were recorded simultaneously from 53 male and 53 female subjects saying sustained /A/, /i/, and /u/ at five equally spaced pitch heights, each being higher than the preceding one. Then analyses of means, variance, and correlation were performed to explore the relationships of VAT/pitch levels and VAT/vowels. Findings were As mean STs (semitone) increase linearly from levels 1 to 5, mean VATs decrease nonlinearly in a big group of subjects but increase nonlinearly in a small group of them. Based on the body-cover model of F0 control, data here lead to the guess that different people incline to use different strategies in increasing pitch height. When males, females, and males plus females are considered as a whole, average STs and VATs tend to be positively correlated among the three vertex vowels.

  15. Tongue movements in vowel-consonant-vowel (VCV) sequences: The effect of consonant length

    NASA Astrophysics Data System (ADS)

    Lofqvist, Anders

    2005-09-01

    This study examined the effect of consonant duration on the tongue movement from the first to the second vowel in VCV sequences, where the consonant is a short or long labial nasal consonant. Lip, tongue, and jaw movements were recorded in native speakers of Japanese using a magnetometer system. Measurements were made of the duration, path, and speed of the tongue movement trajectory between the two vowels. The coordination of the onsets of the lip closing and tongue movements was also studied, as well as the relative part of the trajectory that occurred during the consonant and the vowels. Preliminary results show a robust difference in duration between the long and short consonants, with the long ones about twice as long. The duration of the tongue movement was longer in the long than the short consonants. Both the peak and average speed of the tongue movement were slower in the long consonants. The tongue movement path was slightly longer in the long consonants. These results suggest that speakers adjust the tongue movement trajectory so that a similar relationship between the movement and the consonant closure is maintained in both the long and the short consonants. [Work supported by NIH.

  16. The acoustic and visual factors influencing the construction of tranquil space in urban and rural environments tranquil spaces-quiet places?

    PubMed

    Pheasant, Robert; Horoshenkov, Kirill; Watts, Greg; Barrett, Brendan

    2008-03-01

    Prior to this work no structured mechanism existed in the UK to evaluate the tranquillity of open spaces with respect to the characteristics of both acoustic and visual stimuli. This is largely due to the fact that within the context of "tranquil" environments, little is known about the interaction of the audio-visual modalities and how they combine to lead to the perception of tranquillity. This paper presents the findings of a study in which visual and acoustic data, captured from 11 English rural and urban landscapes, were used by 44 volunteers to make subjective assessments of both their perceived tranquillity of a location, and the loudness of five generic soundscape components. The results were then analyzed alongside objective measurements taken in the laboratory. It was found that the maximum sound pressure level (L(Amax)) and the percentage of natural features present at a location were the key factors influencing tranquillity. Engineering formulas for the tranquillity as a function of the noise level and proportion of the natural features are proposed.

  17. Instrumental Dimensioning of Normal and Pathological Phonation Using Acoustic Measurements

    ERIC Educational Resources Information Center

    Putzer, Manfred; Barry, William J.

    2008-01-01

    The present study deals with the dimensions of normal and pathological phonation. Separation of normal voices from pathological voices is tested under different aspects. Using a new parametrization of voice-quality properties in the acoustic signal, the vowel productions of 534 speakers (267 M, 267 F) without any reported voice pathology and the…

  18. Use of Acoustic Cues by Children with Cochlear Implants

    ERIC Educational Resources Information Center

    Giezen, Marcel R.; Escudero, Paola; Baker, Anne

    2010-01-01

    Purpose: This study examined the use of different acoustic cues in auditory perception of consonant and vowel contrasts by profoundly deaf children with a cochlear implant (CI) in comparison to age-matched children and young adults with normal hearing. Method: A speech sound categorization task in an XAB format was administered to 15 children ages…

  19. Tutorial on architectural acoustics

    NASA Astrophysics Data System (ADS)

    Shaw, Neil; Talaske, Rick; Bistafa, Sylvio

    2002-11-01

    This tutorial is intended to provide an overview of current knowledge and practice in architectural acoustics. Topics covered will include basic concepts and history, acoustics of small rooms (small rooms for speech such as classrooms and meeting rooms, music studios, small critical listening spaces such as home theatres) and the acoustics of large rooms (larger assembly halls, auditoria, and performance halls).

  20. Acoustic telemetry and network analysis reveal the space use of multiple reef predators and enhance marine protected area design.

    PubMed

    Lea, James S E; Humphries, Nicolas E; von Brandis, Rainer G; Clarke, Christopher R; Sims, David W

    2016-07-13

    Marine protected areas (MPAs) are commonly employed to protect ecosystems from threats like overfishing. Ideally, MPA design should incorporate movement data from multiple target species to ensure sufficient habitat is protected. We used long-term acoustic telemetry and network analysis to determine the fine-scale space use of five shark and one turtle species at a remote atoll in the Seychelles, Indian Ocean, and evaluate the efficacy of a proposed MPA. Results revealed strong, species-specific habitat use in both sharks and turtles, with corresponding variation in MPA use. Defining the MPA's boundary from the edge of the reef flat at low tide instead of the beach at high tide (the current best in Seychelles) significantly increased the MPA's coverage of predator movements by an average of 34%. Informed by these results, the larger MPA was adopted by the Seychelles government, demonstrating how telemetry data can improve shark spatial conservation by affecting policy directly. PMID:27412274

  1. Acoustic telemetry and network analysis reveal the space use of multiple reef predators and enhance marine protected area design.

    PubMed

    Lea, James S E; Humphries, Nicolas E; von Brandis, Rainer G; Clarke, Christopher R; Sims, David W

    2016-07-13

    Marine protected areas (MPAs) are commonly employed to protect ecosystems from threats like overfishing. Ideally, MPA design should incorporate movement data from multiple target species to ensure sufficient habitat is protected. We used long-term acoustic telemetry and network analysis to determine the fine-scale space use of five shark and one turtle species at a remote atoll in the Seychelles, Indian Ocean, and evaluate the efficacy of a proposed MPA. Results revealed strong, species-specific habitat use in both sharks and turtles, with corresponding variation in MPA use. Defining the MPA's boundary from the edge of the reef flat at low tide instead of the beach at high tide (the current best in Seychelles) significantly increased the MPA's coverage of predator movements by an average of 34%. Informed by these results, the larger MPA was adopted by the Seychelles government, demonstrating how telemetry data can improve shark spatial conservation by affecting policy directly.

  2. Second vowel formant relationship to adduction: A preliminary study

    NASA Astrophysics Data System (ADS)

    Hanrahan, Kevin G.

    The relationship between the vocal tract and the larynx in the formation of vowels has been debated for decades. Vowels were first thought to have been formed in the larynx; then later it was believed that they were formed solely in the vocal tract. In the 1960s Fant formalized this belief into the Source-Filter Theory of Vowel Formation. The theory was interpreted by voice teachers to mean that the larynx had very little to do with the formation of vowels, and this interpretation has dominated voice teaching for decades. Recent research, however, is now suggesting that the larynx and the vocal tract are interactive with each other, meaning that a change of muscular function in the larynx will create a change of resonator function in the vocal tract, and vice versa. This conclusion is drawn mainly on the work of Titze, Story, Laukkanen, et.al. They have found that a relationship exists between laryngeal function and the first vowel formant (F1). When examining research on the second vowel formant (F2), this author discovered that there may be a relationship between F2 and adduction. Therefore, based on present evidence, it was hypothesized that an elevated frequency of F2 corresponded to an increase in adduction. The hypothesis was examined by comparing the resonance output and glottal closure between vowels where F2 was elevated and vowels without modification of F2. Subjects were asked to sing [i], [a], and [u] at a medium dynamic level on D4, G#4, and D5 for the female subjects and an octave below for the male subjects, once using a "generic" version of the vowel, meaning what they considered a "nice, easy, and generic" version of the vowel to be, and then again modifying the vowel to increase the frequency of the upper harmonics. Electroglottogram, pitch, intensity, and formant data were collected and compared. An increase in the frequency of F2 corresponded to an increase in the Closed Quotient (CQ), the length of time the vocal folds are closed, in a few

  3. Mathematical modeling of vowel perception by users of analog multichannel cochlear implants: temporal and channel-amplitude cues.

    PubMed

    Svirsky, M A

    2000-03-01

    A "multidimensional phoneme identification" (MPI) model is proposed to account for vowel perception by cochlear implant users. A multidimensional extension of the Durlach-Braida model of intensity perception, this model incorporates an internal noise model and a decision model to account separately for errors due to poor sensitivity and response bias. The MPI model provides a complete quantitative description of how listeners encode and combine acoustic cues, and how they use this information to determine which sound they heard. Thus, it allows for testing specific hypotheses about phoneme identification in a very stringent fashion. As an example of the model's application, vowel identification matrices obtained with synthetic speech stimuli (including "conflicting cue" conditions [Dorman et al., J. Acoust. Soc. Am. 92, 3428-3432 (1992)] were examined. The listeners were users of the "compressed-analog" stimulation strategy, which filters the speech spectrum into four partly overlapping frequency bands and delivers each signal to one of four electrodes in the cochlea. It was found that a simple model incorporating one temporal cue (i.e., an acoustic cue based only on the time waveforms delivered to the most basal channel) and spectral cues (based on the distribution of amplitudes among channels) can be quite successful in explaining listener responses. The new approach represented by the MPI model may be used to obtain useful insights about speech perception by cochlear implant users in particular, and by all kinds of listeners in general.

  4. The role of selective attention in the acquisition of English tense and lax vowels by native Spanish listeners: comparison of three training methods

    PubMed Central

    Kondaurova, Maria V.; Francis, Alexander L.

    2010-01-01

    This study investigates the role of two processes, cue enhancement (learning to attend to acoustic cues which characterize a speech contrast for native listeners) and cue inhibition (learning to ignore cues that do not), in the acquisition of the American English tense and lax ([i] vs.[I]) vowels by native Spanish listeners. This contrast is acoustically distinguished by both vowel spectrum and duration. However, while native English listeners rely primarily on spectrum, inexperienced Spanish listeners tend to rely exclusively on duration. Twenty-nine native Spanish listeners, initially reliant on vowel duration, received either enhancement training, inhibition training, or training with a natural cue distribution. Results demonstrated that reliance on spectrum properties increased over baseline for all three groups. However, inhibitory training was more effective relative to enhancement training and both inhibitory and enhancement training were more effective relative to natural distribution training in decreasing listeners’ attention to duration. These results suggest that phonetic learning may involve two distinct cognitive processes, cue enhancement and cue inhibition, that function to shift selective attention between separable acoustic dimensions. Moreover, cue-specific training (whether enhancing or inhibitory) appears to be more effective for the acquisition of second language speech contrasts. PMID:21499531

  5. The acoustics of uvulars in Tlingit

    NASA Astrophysics Data System (ADS)

    Denzer-King, Ryan E.

    This paper looks at the acoustics of uvulars in Tlingit, an Athabaskan language spoken in Alaska and Canada. Data from five native speakers was used for acoustic analysis for tokens from five phoneme groups (alveolars, plain velars, labialized velars, plain uvulars, and labialized uvulars). The tokens were analyzed by computing spectral moments of plosive bursts and fricatives, and F2 and F3 values for post-consonantal vowels, which were used to calculate locus equations, a descriptive measure of the relationship between F2 at vowel onset and midpoint. Several trends were observed, including a greater difference between F2 and F 3 after uvulars than after velars, as well as a higher center of gravity (COG) and lower skew and kurtosis for uvulars than for velars. The comparison of plain versus labialized consonants supports the finding of Suh (2008) that labialization lowers mean burst energy, or COG, and additionally found labialization to raise skew and kurtosis.

  6. Generalization of von Neumann analysis for a model of two discrete half-spaces: The acoustic case

    USGS Publications Warehouse

    Haney, M.M.

    2007-01-01

    Evaluating the performance of finite-difference algorithms typically uses a technique known as von Neumann analysis. For a given algorithm, application of the technique yields both a dispersion relation valid for the discrete time-space grid and a mathematical condition for stability. In practice, a major shortcoming of conventional von Neumann analysis is that it can be applied only to an idealized numerical model - that of an infinite, homogeneous whole space. Experience has shown that numerical instabilities often arise in finite-difference simulations of wave propagation at interfaces with strong material contrasts. These interface instabilities occur even though the conventional von Neumann stability criterion may be satisfied at each point of the numerical model. To address this issue, I generalize von Neumann analysis for a model of two half-spaces. I perform the analysis for the case of acoustic wave propagation using a standard staggered-grid finite-difference numerical scheme. By deriving expressions for the discrete reflection and transmission coefficients, I study under what conditions the discrete reflection and transmission coefficients become unbounded. I find that instabilities encountered in numerical modeling near interfaces with strong material contrasts are linked to these cases and develop a modified stability criterion that takes into account the resulting instabilities. I test and verify the stability criterion by executing a finite-difference algorithm under conditions predicted to be stable and unstable. ?? 2007 Society of Exploration Geophysicists.

  7. Greek perception and production of an English vowel contrast: A preliminary study

    NASA Astrophysics Data System (ADS)

    Podlipský, Václav J.

    2005-04-01

    This study focused on language-independent principles functioning in acquisition of second language (L2) contrasts. Specifically, it tested Bohn's Desensitization Hypothesis [in Speech perception and linguistic experience: Issues in Cross Language Research, edited by W. Strange (York Press, Baltimore, 1995)] which predicted that Greek speakers of English as an L2 would base their perceptual identification of English /i/ and /I/ on durational differences. Synthetic vowels differing orthogonally in duration and spectrum between the /i/ and /I/ endpoints served as stimuli for a forced-choice identification test. To assess L2 proficiency and to evaluate the possibility of cross-language category assimilation, productions of English /i/, /I/, and /ɛ/ and of Greek /i/ and /e/ were elicited and analyzed acoustically. The L2 utterances were also rated for the degree of foreign accent. Two native speakers of Modern Greek with low and 2 with intermediate experience in English participated. Six native English (NE) listeners and 6 NE speakers tested in an earlier study constituted the control groups. Heterogeneous perceptual behavior was observed for the L2 subjects. It is concluded that until acquisition in completely naturalistic settings is tested, possible interference of formally induced meta-linguistic differentiation between a ``short'' and a ``long'' vowel cannot be eliminated.

  8. Effects of Consonant-Vowel Transitions in Speech Stimuli on Cortical Auditory Evoked Potentials in Adults

    PubMed Central

    Doellinger, Michael; Burger, Martin; Hoppe, Ulrich; Bosco, Enrico; Eysholdt, Ulrich

    2011-01-01

    We examined the neural activation to consonant-vowel transitions by cortical auditory evoked potentials (AEPs). The aim was to show whether cortical response patterns to speech stimuli contain components due to one of the temporal features, the voice-onset time (VOT). In seven normal-hearing adults, the cortical responses to four different monosyllabic words were opposed to the cortical responses to noise stimuli with the same temporal envelope as the speech stimuli. Significant hemispheric asymmetries were found for speech but not in noise evoked potentials. The difference signals between the AEPs to speech and corresponding noise stimuli revealed a significant negative component, which correlated with the VOT. The hemispheric asymmetries can be referred to rapid spectral changes. The correlation with the VOT indicates that the significant component in the difference signal reflects the perception of the acoustic change within the consonant-vowel transition. Thus, at the level of automatic processing, the characteristics of speech evoked potentials appear to be determined primarily by temporal aspects of the eliciting stimuli. PMID:21643536

  9. The effect of language immersion education on the preattentive perception of native and non-native vowel contrasts.

    PubMed

    Peltola, Maija S; Tuomainen, Outi; Koskinen, Mira; Aaltonen, Olli

    2007-01-01

    Proficiency in a second language (L2) may depend upon the age of exposure and the continued use of the mother tongue (L1) during L2 acquisition. The effect of early L2 exposure on the preattentive perception of native and non-native vowel contrasts was studied by measuring the mismatch negativity (MMN) response from 14-year-old children. The test group consisted of six Finnish children who had participated in English immersion education. The control group consisted of eight monolingual Finns. The subjects were presented with Finnish and English synthetic vowel contrasts. The aim was to see whether early exposure had resulted in the development of a new language-specific memory trace for the contrast phonemically irrelevant in L1. The results indicated that only the contrast with the largest acoustic distance elicited an MMN response in the Bilingual group, while the Monolingual group showed a response also to the native contrast. This may suggest that native-like memory traces for prototypical vowels were not formed in early language immersion. PMID:17109242

  10. The Basis for Language Acquisition: Congenitally Deaf Infants Discriminate Vowel Length in the First Months after Cochlear Implantation.

    PubMed

    Vavatzanidis, Niki Katerina; Mürbe, Dirk; Friederici, Angela; Hahne, Anja

    2015-12-01

    One main incentive for supplying hearing impaired children with a cochlear implant is the prospect of oral language acquisition. Only scarce knowledge exists, however, of what congenitally deaf children actually perceive when receiving their first auditory input, and specifically what speech-relevant features they are able to extract from the new modality. We therefore presented congenitally deaf infants and young children implanted before the age of 4 years with an oddball paradigm of long and short vowel variants of the syllable /ba/. We measured the EEG in regular intervals to study their discriminative ability starting with the first activation of the implant up to 8 months later. We were thus able to time-track the emerging ability to differentiate one of the most basic linguistic features that bears semantic differentiation and helps in word segmentation, namely, vowel length. Results show that already 2 months after the first auditory input, but not directly after implant activation, these early implanted children differentiate between long and short syllables. Surprisingly, after only 4 months of hearing experience, the ERPs have reached the same properties as those of the normal hearing control group, demonstrating the plasticity of the brain with respect to the new modality. We thus show that a simple but linguistically highly relevant feature such as vowel length reaches age-appropriate electrophysiological levels as fast as 4 months after the first acoustic stimulation, providing an important basis for further language acquisition.

  11. Early learners' discrimination of second-language vowels.

    PubMed

    Højen, Anders; Flege, James E

    2006-05-01

    It is uncertain from previous research to what extent the perceptual system retains plasticity after attunement to the native language (L1) sound system. This study evaluated second-language (L2) vowel discrimination by individuals who began learning the L2 as children ("early learners"). Experiment 1 identified procedures that lowered discrimination scores for foreign vowel contrasts in an AXB test (with three physically different stimuli per trial, where "X" was drawn from the same vowel category as "A" or "B"). Experiment 2 examined the AXB discrimination of English vowels by native Spanish early learners and monolingual speakers of Spanish and English (20 per group) at interstimulus intervals (ISIs) of 1000 and 0 ms. The Spanish monolinguals obtained near-chance scores for three difficult vowel contrasts, presumably because they did not perceive the vowels as distinct phonemes and because the experimental design hindered low-level encoding strategies. Like the English monolinguals, the early learners obtained high scores, indicating they had shown considerable perceptual learning. However, statistically significant differences between early learners and English monolinguals for two of three difficult contrasts at the 0-ms ISI suggested that their underlying perceptual systems were not identical. Implications for claims regarding perceptual plasticity following L1 attunement are discussed. PMID:16708962

  12. Nonlinear Dust Acoustic Waves in Dissipative Space Dusty Plasmas with Superthermal Electrons and Nonextensive Ions

    NASA Astrophysics Data System (ADS)

    El-Hanbaly, A. M.; El-Shewy, E. K.; Sallah, M.; Darweesh, H. F.

    2016-05-01

    The nonlinear characteristics of the dust acoustic (DA) waves are studied in a homogeneous, collisionless, unmagnetized, and dissipative dusty plasma composed of negatively charged dusty grains, superthermal electrons, and nonextensive ions. Sagdeev pseudopotential technique has been employed to study the large amplitude DA waves. It (Sagdeev pseudopotential) has an evidence for the existence of compressive and rarefractive solitons. The global features of the phase portrait are investigated to understand the possible types of solutions of the Sagdeev form. On the other hand, the reductive perturbation technique has been used to study small amplitude DA waves and yields the Korteweg-de Vries-Burgers (KdV-Burgers) equation that exhibits both soliton and shock waves. The behavior of the obtained results of both large and small amplitude is investigated graphically in terms of the plasma parameters like dust kinematic viscosity, superthermal and nonextensive parameters.

  13. Acoustic emission frequency discrimination

    NASA Technical Reports Server (NTRS)

    Sugg, Frank E. (Inventor); Graham, Lloyd J. (Inventor)

    1988-01-01

    In acoustic emission nondestructive testing, broadband frequency noise is distinguished from narrow banded acoustic emission signals, since the latter are valid events indicative of structural flaws in the material being examined. This is accomplished by separating out those signals which contain frequency components both within and beyond (either above or below) the range of valid acoustic emission events. Application to acoustic emission monitoring during nondestructive bond verification and proof loading of undensified tiles on the Space Shuttle Orbiter is considered.

  14. An evaluation of Space Shuttle STS-2 payload bay acoustic data and comparison with predictions

    NASA Technical Reports Server (NTRS)

    Wilby, J. F.; Piersol, A. G.; Wilby, E. G.

    1982-01-01

    Space average sound pressure levels computed from measurements at 18 locations in the payload bay of the Space Shuttle orbiter vehicle during the STS-2 launch were compared with predicted levels obtained using the PACES computer program. The comparisons were performed over the frequency range 12.5 Hz to 1000 Hz, since the test data at higher frequencies are contaminated by instrumentation background noise. In general the PACES computer program tends to overpredict the space average sound levels in the payload bay, although the magnitude of the discrepancy is usually small. Furthermore the discrepancy depends to some extent on the manner in which the payload is modeled analytically, and the method used to determine the "measured' space average sound pressure levels. Thus the difference between predicted and measured sound levels, averaged over the 20 one third octave bands from 12.5 Hz to 1000 Hz, varies from 1 dB to 3.5 dB.

  15. A dedicated neural mechanism for vowel selection: a case of relative vowel deficit sparing the number lexicon.

    PubMed

    Semenza, Carlo; Bencini, Giulia M L; Bertella, Laura; Mori, Ileana; Pignatti, Riccardo; Ceriani, Francesca; Cherrick, Danielle; Caldognetto, Emanuela Magno

    2007-01-28

    We report the case of an Italian speaker (GBC), with classical Wernicke's aphasia following a vascular lesion in the posterior middle temporal region. GBC exhibited a selective deficit in spoken language production affecting vowels more than consonants. In reading from a newspaper, GBC substituted vowels for other vowels from the Italian inventory at a rate of 7/1 compared to consonants. No effects of frequency or grammatical class were found. Vowel errors could also not be accounted for by morphological or known phonological processes. Production of number words, in contrast, was free from phonological errors. While GBC has intact representations of Italian vowels and consonants, his data argue for a separate selection mechanism for vowels that is dissociable from that used for consonants. This case provides neuropsychological evidence for models of word production that distinguish between the abstract phonological properties of a word (e.g., sequencing of phonemic slots, or "CV skeleton") and a separate representation for the specific sounds (melody). PMID:16997332

  16. Infant cortical electrophysiology and perception of vowel contrasts.

    PubMed

    Cone, Barbara K

    2015-02-01

    Cortical auditory evoked potentials (CAEPs) were obtained for vowel tokens presented in an oddball stimulus paradigm. Perceptual measures of vowel discrimination were obtained using a visually-reinforced head-turn paradigm. The hypothesis was that CAEP latencies and amplitudes would differ as a function of vowel type and be correlated with perceptual performance. Twenty normally hearing infants aged 4-12 months were evaluated. CAEP component amplitudes and latencies were measured in response to the standard, frequent token /a/ and for infrequent, deviant tokens /i/, /o/ and /u/, presented at rates of 1 and 2 tokens/s. The perceptual task required infants to make a behavioral response for trials that contained two different vowel tokens, and ignore those in which the tokens were the same. CAEP amplitudes were larger in response to the deviant tokens, when compared to the control condition in which /a/ served as both standard and deviant. This was also seen in waveforms derived by subtracting the response to standard /a/ from the responses to deviant tokens. CAEP component latencies in derived responses at 2/s also demonstrated some sensitivity to vowel contrast type. The average hit rate for the perceptual task was 68.5%, with a 25.7% false alarm rate. There were modest correlations of CAEP amplitudes and latencies with perceptual performance. The CAEP amplitude differences for vowel contrasts could be used as an indicator of the underlying neural capacity to encode spectro-temporal differences in vowel sounds. This technique holds promise for translation to clinical methods for evaluating speech perception.

  17. Infant Cortical Electrophysiology and Perception of Vowel Contrasts

    PubMed Central

    Cone, Barbara K

    2014-01-01

    Cortical auditory evoked potentials (CAEPs) were obtained for vowel tokens presented in an oddball stimulus paradigm. Perceptual measures of vowel discrimination were obtained using a visually-reinforced head-turn paradigm. The hypothesis was that CAEP latencies and amplitudes would differ as a function of vowel type and be correlated with perceptual performance. Twenty normally hearing infants aged 4–12 months were evaluated. CAEP component amplitudes and latencies were measured in response to the standard, frequent token /a/ and for infrequent, deviant tokens /i/, /o/ and /u/, presented at rates of 1 and 2 tokens/second. The perceptual task required infants to make a behavioral response for trials that contained two different vowel tokens, and ignore those in which the tokens were the same. CAEP amplitudes were larger in response to the deviant tokens, when compared to the control condition in which /a/ served as both standard and deviant. This was also seen in waveforms derived by subtracting the response to standard /a/ from the responses to deviant tokens. CAEP component latencies in derived responses at 2/sec also demonstrated some sensitivity to vowel contrast type. The average hit rate for the perceptual task was 68.5%, with a 25.7% false alarm rate. There were modest correlations of CAEP amplitudes and latencies with perceptual performance. The CAEP amplitude differences for vowel contrasts could be used as an indicator of the underlying neural capacity to encode spectro-temporal differences in vowel sounds. This technique holds promise for translation to clinical methods for evaluating speech perception. PMID:24933411

  18. Perception of complete and incomplete formant transitions in vowels

    PubMed Central

    Divenyi, Pierre

    2009-01-01

    In everyday speech, formant transitions rarely reach the canonical frequencies of a target vowel. The perceptual system often compensates for such production undershoots, called vowel reduction (VR), by a perceptual overshoot of the final transition frequencies. The present investigation explored the perceptual parameters and existence region of VR. In a series of experiments a 100-ms steady-state vowel V1 was followed by a formant transition toward a target vowel V2. By manipulating both its duration and velocity, in most stimuli the transition was truncated and only seldom reached the target. After being presented with the vowel V2 before each block of trials, listeners were asked to rate their confidence that the transition actually reached the V2 target. Transitions along six trajectories connecting the three cardinal vowels ∕a/, ∕i/, and ∕u/ in both directions as well as the transition ∕ie∕ (halfway along the trajectory ∕ia∕) were examined in experiments in which either the duration of the transition was fixed and its velocity was varied or vice-versa. Results confirmed the existence of perceptual overshoot and showed that, at the point a transition short of reaching the vowel V2 was just perceived as if it had reached the target, transition duration and transition velocity were inversely related. The amount of overshoot was found to be larger for larger V1-V2 distances and shorter trajectory durations. The overshoot could be reliably predicted by a linear model based on three parameters—the extent of V1-V2 distance, transition velocity, and transition acceleration. These findings suggest that the perceptual dynamics of speech relies on mechanisms that estimate the rate of change in the resonant characteristics of the vocal tract. PMID:19739756

  19. Analysis of Acoustic Features in Speakers with Cognitive Disorders and Speech Impairments

    NASA Astrophysics Data System (ADS)

    Saz, Oscar; Simón, Javier; Rodríguez, W. Ricardo; Lleida, Eduardo; Vaquero, Carlos

    2009-12-01

    This work presents the results in the analysis of the acoustic features (formants and the three suprasegmental features: tone, intensity and duration) of the vowel production in a group of 14 young speakers suffering different kinds of speech impairments due to physical and cognitive disorders. A corpus with unimpaired children's speech is used to determine the reference values for these features in speakers without any kind of speech impairment within the same domain of the impaired speakers; this is 57 isolated words. The signal processing to extract the formant and pitch values is based on a Linear Prediction Coefficients (LPCs) analysis of the segments considered as vowels in a Hidden Markov Model (HMM) based Viterbi forced alignment. Intensity and duration are also based in the outcome of the automated segmentation. As main conclusion of the work, it is shown that intelligibility of the vowel production is lowered in impaired speakers even when the vowel is perceived as correct by human labelers. The decrease in intelligibility is due to a 30% of increase in confusability in the formants map, a reduction of 50% in the discriminative power in energy between stressed and unstressed vowels and to a 50% increase of the standard deviation in the length of the vowels. On the other hand, impaired speakers keep good control of tone in the production of stressed and unstressed vowels.

  20. Entendre, lire et prononcer les voyelles nasales (Hearing, Reading, and Pronouncing Nasal Vowels).

    ERIC Educational Resources Information Center

    Noir, Pascal; Peifer, Michel

    1996-01-01

    Describes techniques used for teaching French language students to recognize, discriminate between, and pronounce the four French nasal vowels. Includes twenty tongue-twisting sentences exemplifying the different vowels. (MSE)

  1. Reading Arabic Texts: Effects of Text Type, Reader Type and Vowelization.

    ERIC Educational Resources Information Center

    Abu-Rabia, Salim

    1998-01-01

    Investigates the effect of vowels on reading accuracy in Arabic orthography. Finds that vowels had a significant effect on reading accuracy of poor and skilled readers in reading each of four kinds of texts. (NH)

  2. Quantifying loss of acoustic communication space for right whales in and around a U.S. National Marine Sanctuary.

    PubMed

    Hatch, Leila T; Clark, Christopher W; Van Parijs, Sofie M; Frankel, Adam S; Ponirakis, Dimitri W

    2012-12-01

    The effects of chronic exposure to increasing levels of human-induced underwater noise on marine animal populations reliant on sound for communication are poorly understood. We sought to further develop methods of quantifying the effects of communication masking associated with human-induced sound on contact-calling North Atlantic right whales (Eubalaena glacialis) in an ecologically relevant area (~10,000 km(2) ) and time period (peak feeding time). We used an array of temporary, bottom-mounted, autonomous acoustic recorders in the Stellwagen Bank National Marine Sanctuary to monitor ambient noise levels, measure levels of sound associated with vessels, and detect and locate calling whales. We related wind speed, as recorded by regional oceanographic buoys, to ambient noise levels. We used vessel-tracking data from the Automatic Identification System to quantify acoustic signatures of large commercial vessels. On the basis of these integrated sound fields, median signal excess (the difference between the signal-to-noise ratio and the assumed recognition differential) for contact-calling right whales was negative (-1 dB) under current ambient noise levels and was further reduced (-2 dB) by the addition of noise from ships. Compared with potential communication space available under historically lower noise conditions, calling right whales may have lost, on average, 63-67% of their communication space. One or more of the 89 calling whales in the study area was exposed to noise levels ≥120 dB re 1 μPa by ships for 20% of the month, and a maximum of 11 whales were exposed to noise at or above this level during a single 10-min period. These results highlight the limitations of exposure-threshold (i.e., dose-response) metrics for assessing chronic anthropogenic noise effects on communication opportunities. Our methods can be used to integrate chronic and wide-ranging noise effects in emerging ocean-planning forums that seek to improve management of cumulative effects

  3. Structural-acoustic optimum design of shell structures in open/closed space based on a free-form optimization method

    NASA Astrophysics Data System (ADS)

    Shimoda, Masatoshi; Shimoide, Kensuke; Shi, Jin-Xing

    2016-03-01

    Noise reduction by structural geometry optimization has attracted much attention among designers. In the present work, we propose a free-form optimization method for the structural-acoustic design optimization of shell structures to reduce the noise of a targeted frequency or frequency range in an open or closed space. The objective of the design optimization is to minimize the average structural vibration-induced sound pressure at the evaluation points in the acoustic field under a volume constraint. For the shape design optimization, we carry out structural-acoustic coupling analysis and adjoint analysis to calculate the shape gradient functions. Then, we use the shape gradient functions in velocity analysis to update the shape of shell structures. We repeat this process until convergence is confirmed to obtain the optimum shape of the shell structures in a structural-acoustic coupling system. The numerical results for the considered examples showed that the proposed design optimization process can significantly reduce the noise in both open and closed spaces.

  4. Acoustic source for generating an acoustic beam

    DOEpatents

    Vu, Cung Khac; Sinha, Dipen N.; Pantea, Cristian

    2016-05-31

    An acoustic source for generating an acoustic beam includes a housing; a plurality of spaced apart piezo-electric layers disposed within the housing; and a non-linear medium filling between the plurality of layers. Each of the plurality of piezoelectric layers is configured to generate an acoustic wave. The non-linear medium and the plurality of piezo-electric material layers have a matching impedance so as to enhance a transmission of the acoustic wave generated by each of plurality of layers through the remaining plurality of layers.

  5. Vowel Targeted Intervention for Children with Persisting Speech Difficulties: Impact on Intelligibility

    ERIC Educational Resources Information Center

    Speake, Jane; Stackhouse, Joy; Pascoe, Michelle

    2012-01-01

    Compared to the treatment of consonant segments, the treatment of vowels is infrequently described in the literature on children's speech difficulties. Vowel difficulties occur less frequently than those with consonants but may have significant impact on intelligibility. In order to evaluate the effectiveness of vowel targeted intervention (VTI)…

  6. Structural Generalizations over Consonants and Vowels in 11-Month-Old Infants

    ERIC Educational Resources Information Center

    Pons, Ferran; Toro, Juan M.

    2010-01-01

    Recent research has suggested consonants and vowels serve different roles during language processing. While statistical computations are preferentially made over consonants but not over vowels, simple structural generalizations are easily made over vowels but not over consonants. Nevertheless, the origins of this asymmetry are unknown. Here we…

  7. Improving L2 Listeners' Perception of English Vowels: A Computer-Mediated Approach

    ERIC Educational Resources Information Center

    Thomson, Ron I.

    2012-01-01

    A high variability phonetic training technique was employed to train 26 Mandarin speakers to better perceive ten English vowels. In eight short training sessions, learners identified 200 English vowel tokens, produced in a post bilabial stop context by 20 native speakers. Learners' ability to identify English vowels significantly improved in the…

  8. Biomechanically Preferred Consonant-Vowel Combinations Fail to Appear in Adult Spoken Corpora

    ERIC Educational Resources Information Center

    Whalen, D. H.; Giulivi, Sara; Nam, Hosung; Levitt, Andrea G.; Halle, Pierre; Goldstein, Louis M.

    2012-01-01

    Certain consonant/vowel (CV) combinations are more frequent than would be expected from the individual C and V frequencies alone, both in babbling and, to a lesser extent, in adult language, based on dictionary counts: Labial consonants co-occur with central vowels more often than chance would dictate; coronals co-occur with front vowels, and…

  9. Vowel Cluster-Phoneme Correspondences in 20,000 English Words.

    ERIC Educational Resources Information Center

    Johnson, Dale D.

    The symbol-sound correspondence status of vowel-cluster (two or more adjacent vowel letters) spelling in American English was investigated. The source of the study was Venezky's 1963 revision of the Thorndike Frequency Count. A computer print-out of the 20,000 word corpus was analyzed to determine the letter-sound correspondence of vowel cluster…

  10. Vowel Patterns in Developmental Apraxia of Speech: Three Longitudinal Case Studies

    ERIC Educational Resources Information Center

    Davis, Barbara L.; Jacks, Adam; Marquardt, Thomas P.

    2005-01-01

    Vowel inventories and error patterns for three children with suspected developmental apraxia of speech (DAS) were analysed over a 3-year period using phonetic transcriptions of connected speech samples. The children demonstrated complete English vowel inventories except for rhotics. However, accuracy of vowel targets in connected speech did not…

  11. Textual Input Enhancement for Vowel Blindness: A Study with Arabic ESL Learners

    ERIC Educational Resources Information Center

    Alsadoon, Reem; Heift, Trude

    2015-01-01

    This study explores the impact of textual input enhancement on the noticing and intake of English vowels by Arabic L2 learners of English. Arabic L1 speakers are known to experience "vowel blindness," commonly defined as a difficulty in the textual decoding and encoding of English vowels due to an insufficient decoding of the word form.…

  12. The Role of Vowels in Reading Semitic Scripts: Data from Arabic and Hebrew.

    ERIC Educational Resources Information Center

    Abu-Rabia, Salim

    2001-01-01

    Investigates the effect of vowels and context on reading accuracy of skilled adult native Arabic speakers in Arabic and in Hebrew, their second language. Reveals a significant effect for vowels and for context across all reading conditions in Arabic and Hebrew. Finds that the vowelized texts in Arabic and the pointed and unpointed texts in Hebrew…

  13. Speech recognition: Acoustic, phonetic and lexical knowledge

    NASA Astrophysics Data System (ADS)

    Zue, V. W.

    1985-08-01

    During this reporting period we continued to make progress on the acquisition of acoustic-phonetic and lexical knowledge. We completed development of a continuous digit recognition system. The system was constructed to investigate the use of acoustic-phonetic knowledge in a speech recognition system. The significant achievements of this study include the development of a soft-failure procedure for lexical access and the discovery of a set of acoustic-phonetic features for verification. We completed a study of the constraints that lexical stress imposes on word recognition. We found that lexical stress information alone can, on the average, reduce the number of word candidates from a large dictionary by more than 80 percent. In conjunction with this study, we successfully developed a system that automatically determines the stress pattern of a word from the acoustic signal. We performed an acoustic study on the characteristics of nasal consonants and nasalized vowels. We have also developed recognition algorithms for nasal murmurs and nasalized vowels in continuous speech. We finished the preliminary development of a system that aligns a speech waveform with the corresponding phonetic transcription.

  14. Asymmetries in the Processing of Vowel Height

    ERIC Educational Resources Information Center

    Scharinger, Mathias; Monahan, Philip J.; Idsardi, William J.

    2012-01-01

    Purpose: Speech perception can be described as the transformation of continuous acoustic information into discrete memory representations. Therefore, research on neural representations of speech sounds is particularly important for a better understanding of this transformation. Speech perception models make specific assumptions regarding the…

  15. Acoustics and perception of overtone singing.

    PubMed

    Bloothooft, G; Bringmann, E; van Cappellen, M; van Luipen, J B; Thomassen, K P

    1992-10-01

    Overtone singing, a technique of Asian origin, is a special type of voice production resulting in a very pronounced, high and separate tone that can be heard over a more or less constant drone. An acoustic analysis is presented of the phenomenon and the results are described in terms of the classical theory of speech production. The overtone sound may be interpreted as the result of an interaction of closely spaced formants. For the lower overtones, these may be the first and second formant, separated from the lower harmonics by a nasal pole-zero pair, as the result of a nasalized articulation shifting from /c/ to /a/, or, as an alternative, the second formant alone, separated from the first formant by the nasal pole-zero pair, again as the result of a nasalized articulation around /c/. For overtones with a frequency higher than 800 Hz, the overtone sound can be explained as a combination of the second and third formant as the result of a careful, retroflex, and rounded articulation from /c/, via schwa /e/ to /y/ and /i/ for the highest overtones. The results indicate a firm and relatively long closure of the glottis during overtone phonation. The corresponding short open duration of the glottis introduces a glottal formant that may enhance the amplitude of the intended overtone. Perception experiments showed that listeners categorized the overtone sounds differently from normally sung vowels, which possibly has its basis in an independent perception of the small bandwidth of the resonance underlying the overtone. Their verbal judgments were in agreement with the presented phonetic-acoustic explanation.

  16. Impact of chevron spacing and asymmetric distribution on supersonic jet acoustics and flow

    NASA Astrophysics Data System (ADS)

    Heeb, N.; Gutmark, E.; Kailasanath, K.

    2016-05-01

    An experimental investigation into the effect of chevron spacing and distribution on supersonic jets was performed. Cross-stream and streamwise particle imaging velocimetry measurements were used to relate flow field modification to sound field changes measured by far-field microphones in the overexpanded, ideally expanded, and underexpanded regimes. Drastic modification of the jet cross-section was achieved by the investigated configurations, with both elliptic and triangular shapes attained downstream. Consequently, screech was nearly eliminated with reductions in the range of 10-25 dB depending on the operating condition. Analysis of the streamwise velocity indicated that both the mean shock spacing and strength were reduced resulting in an increase in the broadband shock associated noise spectral peak frequency and a reduction in the amplitude, respectively. Maximum broadband shock associated noise amplitude reductions were in the 5-7 dB range. Chevron proximity was found to be the primary driver of peak vorticity production, though persistence followed the opposite trend. The integrated streamwise vorticity modulus was found to be correlated with peak large scale turbulent mixing noise reduction, though optimal overall sound pressure level reductions did not necessarily follow due to the shock/fine scale mixing noise sources. Optimal large scale mixing noise reductions were in the 5-6 dB range.

  17. Voiceless Bangla vowel recognition using sEMG signal.

    PubMed

    Mostafa, S S; Awal, M A; Ahmad, M; Rashid, M A

    2016-01-01

    Some people cannot produce sound although their facial muscles work properly due to having problem in their vocal cords. Therefore, recognition of alphabets as well as sentences uttered by these voiceless people is a complex task. This paper proposes a novel method to solve this problem using non-invasive surface Electromyogram (sEMG). Firstly, eleven Bangla vowels are pronounced and sEMG signals are recorded at the same time. Different features are extracted and mRMR feature selection algorithm is then applied to select prominent feature subset from the large feature vector. After that, these prominent features subset is applied in the Artificial Neural Network for vowel classification. This novel Bangla vowel classification method can offer a significant contribution in voice synthesis as well as in speech communication. The result of this experiment shows an overall accuracy of 82.3 % with fewer features compared to other studies in different languages. PMID:27652095

  18. Hopfield model applied to vowel and consonant discrimination

    NASA Astrophysics Data System (ADS)

    Gold, Bernard

    1986-08-01

    A Hopfield model of 120 ``neurons'' has been simulated on an LDSP (Lincoln Digital Signal Processor). The model has been applied to the study of problems in automatic discrimination of vowels and consonants. In the first problem, a spectral cross section was extracted by performing a fast Fourier transform on a 20 ms segment from the steady state portion of the vowel in a single syllable word. The spectrum was then smoothed and a one-bit gradient measure applied at 120 frequency values, thus obtaining an assigned state for that vowel. This procedure was repeated until eight such assigned states were obtained. From this data, the connection matrix Tij was obtained using the associative equation, Tij=Σ7s=0x̂six̂sj Each x̂si was the component of one of the assigned states.

  19. Kinematic dust viscosity effect on linear and nonlinear dust-acoustic waves in space dusty plasmas with nonthermal ions

    NASA Astrophysics Data System (ADS)

    El-Hanbaly, A. M.; Sallah, M.; El-Shewy, E. K.; Darweesh, H. F.

    2015-10-01

    Linear and nonlinear dust-acoustic (DA) waves are studied in a collisionless, unmagnetized and dissipative dusty plasma consisting of negatively charged dust grains, Boltzmann-distributed electrons, and nonthermal ions. The normal mode analysis is used to obtain a linear dispersion relation illustrating the dependence of the wave damping rate on the carrier wave number, the dust viscosity coefficient, the ratio of the ion temperature to the electron temperatures, and the nonthermal parameter. The plasma system is analyzed nonlinearly via the reductive perturbation method that gives the KdV-Burgers equation. Some interesting physical solutions are obtained to study the nonlinear waves. These solutions are related to soliton, a combination between a shock and a soliton, and monotonic and oscillatory shock waves. Their behaviors are illustrated and shown graphically. The characteristics of the DA solitary and shock waves are significantly modified by the presence of nonthermal (fast) ions, the ratio of the ion temperature to the electron temperature, and the dust kinematic viscosity. The topology of the phase portrait and the potential diagram of the KdV-Burgers equation is illustrated, whose advantage is the ability to predict different classes of traveling wave solutions according to different phase orbits. The energy of the soliton wave and the electric field are calculated. The results in this paper can be generalized to analyze the nature of plasma waves in both space and laboratory plasma systems.

  20. Kinematic dust viscosity effect on linear and nonlinear dust-acoustic waves in space dusty plasmas with nonthermal ions

    SciTech Connect

    El-Hanbaly, A. M.; Sallah, M.; El-Shewy, E. K.; Darweesh, H. F.

    2015-10-15

    Linear and nonlinear dust-acoustic (DA) waves are studied in a collisionless, unmagnetized and dissipative dusty plasma consisting of negatively charged dust grains, Boltzmann-distributed electrons, and nonthermal ions. The normal mode analysis is used to obtain a linear dispersion relation illustrating the dependence of the wave damping rate on the carrier wave number, the dust viscosity coefficient, the ratio of the ion temperature to the electron temperatures, and the nonthermal parameter. The plasma system is analyzed nonlinearly via the reductive perturbation method that gives the KdV-Burgers equation. Some interesting physical solutions are obtained to study the nonlinear waves. These solutions are related to soliton, a combination between a shock and a soliton, and monotonic and oscillatory shock waves. Their behaviors are illustrated and shown graphically. The characteristics of the DA solitary and shock waves are significantly modified by the presence of nonthermal (fast) ions, the ratio of the ion temperature to the electron temperature, and the dust kinematic viscosity. The topology of the phase portrait and the potential diagram of the KdV-Burgers equation is illustrated, whose advantage is the ability to predict different classes of traveling wave solutions according to different phase orbits. The energy of the soliton wave and the electric field are calculated. The results in this paper can be generalized to analyze the nature of plasma waves in both space and laboratory plasma systems.

  1. Near noise field characteristics of Nike rocket motors for application to space vehicle payload acoustic qualification

    NASA Technical Reports Server (NTRS)

    Hilton, D. A.; Bruton, D.

    1977-01-01

    Results of a series of noise measurements that were made under controlled conditions during the static firing of two Nike solid propellant rocket motors are presented. The usefulness of these motors as sources for general spacecraft noise testing was assessed, and the noise expected in the cargo bay of the orbiter was reproduced. Brief descriptions of the Nike motor, the general procedures utilized for the noise tests, and representative noise data including overall sound pressure levels, one third octave band spectra, and octave band spectra were reviewed. Data are presented on two motors of different ages in order to show the similarity between noise measurements made on motors having different loading dates. The measured noise from these tests is then compared to that estimated for the space shuttle orbiter cargo bay.

  2. Estimating pore-space gas hydrate saturations from well log acoustic data

    USGS Publications Warehouse

    Lee, Myung W.; Waite, William F.

    2008-01-01

    Relating pore-space gas hydrate saturation to sonic velocity data is important for remotely estimating gas hydrate concentration in sediment. In the present study, sonic velocities of gas hydrate–bearing sands are modeled using a three-phase Biot-type theory in which sand, gas hydrate, and pore fluid form three homogeneous, interwoven frameworks. This theory is developed using well log compressional and shear wave velocity data from the Mallik 5L-38 permafrost gas hydrate research well in Canada and applied to well log data from hydrate-bearing sands in the Alaskan permafrost, Gulf of Mexico, and northern Cascadia margin. Velocity-based gas hydrate saturation estimates are in good agreement with Nuclear Magneto Resonance and resistivity log estimates over the complete range of observed gas hydrate saturations.

  3. Exploring the Acoustic Parameter Space in Ultrasound Therapy: Defining the Threshold for Cavitational Effects

    NASA Astrophysics Data System (ADS)

    Kieran, Kathleen; Hall, Timothy L.; Parsons, Jessica E.; Wolf, J. Stuart; Fowlkes, J. Brian; Cain, Charles A.; Roberts, William W.

    2007-05-01

    Focused ultrasound energy is capable of noninvasively, nonthermally ablating tissue. However, the relative contributions of thermal and cavitational effects in the therapeutic use of ultrasound are poorly understood. We sought to identify the ultrasound parameter space within which tissue can be ablated by solely mechanical means (cavitation), without a significant thermal component. Methods: Ultrasound energy (750 kHz, 20 microsecond pulses) was applied sequentially in a 3×3 grid configuration to the cortical tissue of ex vivo porcine kidneys submerged in degassed water. While maintaining constant energy density, intensity (0.11-211 kW/cm2) and duty cycle (0.04%-CW) were varied widely. A thermocouple co-localized with the center of each grid provided continuous temperature measurements. Following ablations, the kidneys were examined grossly and histologically. Results: Ablated tissue was classified into one of four discrete morphologic categories: blanched (firm, pale, desiccated tissue), disrupted (cavity containing thin, isochromatic liquid; no blanching), mixed blanched/disrupted (cavity containing pale, thick liquid; minimal blanching), and no grossly visible effect. Morphologically similar lesions clustered together within the ultrasound parameter space. Disrupted lesions had significantly lower maximal temperatures (44.2 °C) than desiccated (67.5 °C; p<0.0001) or mixed (59.4 °C; p<0.0001) lesions. Conclusions: In an ex vivo model, we have defined the ultrasound parameters within which mechanical tissue ablation, with minimal thermal components, is possible. Future research in vivo is directed toward optimizing the parameters for cavitational tissue ablation, and better understanding the impact of tissue perfusion on lesion generation and intralesional temperature rise.

  4. Analysis and analogy in the perception of vowels.

    PubMed

    Remez, Robert E; Fellowes, Jennifer M; Blumenthal, Eva Y; Nagel, Dalia Shoretz

    2003-10-01

    In two experiments, we investigated the creation of conceptual analogies to a contrast between vowels. An ordering procedure was used to determine the reliability of simple sensory and abstract analogies to vowel contrasts composed by naive volunteers. The results indicate that test subjects compose stable and consistent analogies to a meaningless segmental linguistic contrast, some invoking simple and complex relational properties. Although in the literature of psychophysics such facility has been explained as an effect of sensory analysis, the present studies indicate the action of a far subtler and more versatile cognitive function akin to the creation of meaning in figurative language. PMID:14704027

  5. Acoustic characteristics of Korean stops in Korean child-directed speech

    NASA Astrophysics Data System (ADS)

    Kim, Minjung; Stoel-Gammon, Carol

    2005-04-01

    A variety of cross-linguistic studies have documented that the acoustic properties of speech addressed to young children include exaggeration of pitch contours and acoustically salient features of phonetic units. It has been suggested that phonetic modifications of child-directed speech facilitate young children's speech perception by providing detailed phonetic information about the target word. While there are several studies reporting vowel modifications in speech to infants (i.e., hyper-articulated vowels), there has been relatively little research about consonant modifications in speech to young children (except for VOT). The present study examines acoustic properties of Korean stops in Korean mothers' speech to their children aged 29 to 38 months (N=6). Korean tense, lax, and aspirated stops are all voiceless in word-initial position, and are perceptually differentiated by several acoustic parameters including VOT, f0 of the following vowel, and the amplitude difference of the first and second harmonics at the voice onset of the following vowel. This study compares values of these parameters in Korean motherese to those in speech to adult Koreans from same speakers. Results focus on the acoustic properties of Korean stops in child-directed speech and how they are modified to help Korean young children learn the three-way phonetic contrast.

  6. Intelligibility of American English Vowels of Native and Non-Native Speakers in Quiet and Speech-Shaped Noise

    ERIC Educational Resources Information Center

    Liu, Chang; Jin, Su-Hyun

    2013-01-01

    This study examined intelligibility of twelve American English vowels produced by English, Chinese, and Korean native speakers in quiet and speech-shaped noise in which vowels were presented at six sensation levels from 0 dB to 10 dB. The slopes of vowel intelligibility functions and the processing time for listeners to identify vowels were…

  7. Native dialect matters: perceptual assimilation of Dutch vowels by Czech listeners.

    PubMed

    Chládková, Kateřina; Podlipský, Václav Jonáš

    2011-10-01

    Naive listeners' perceptual assimilations of non-native vowels to first-language (L1) categories can predict difficulties in the acquisition of second-language vowel systems. This study demonstrates that listeners having two slightly different dialects as their L1s can differ in the perception of foreign vowels. Specifically, the study shows that Bohemian Czech and Moravian Czech listeners assimilate Dutch high front vowels differently to L1 categories. Consequently, the listeners are predicted to follow different paths in acquiring these Dutch vowels. These findings underscore the importance of carefully considering the specific dialect background of participants in foreign- and second-language speech perception studies.

  8. Who's talking now? Infants' perception of vowels with infant vocal properties.

    PubMed

    Polka, Linda; Masapollo, Matthew; Ménard, Lucie

    2014-07-01

    Little is known about infants' abilities to perceive and categorize their own speech sounds or vocalizations produced by other infants. In the present study, prebabbling infants were habituated to /i/ ("ee") or /a/ ("ah") vowels synthesized to simulate men, women, and children, and then were presented with new instances of the habituation vowel and a contrasting vowel on different trials, with all vowels simulating infant talkers. Infants showed greater recovery of interest to the contrasting vowel than to the habituation vowel, which demonstrates recognition of the habituation-vowel category when it was produced by an infant. A second experiment showed that encoding the vowel category and detecting the novel vowel required additional processing when infant vowels were included in the habituation set. Despite these added cognitive demands, infants demonstrated the ability to track vowel categories in a multitalker array that included infant talkers. These findings raise the possibility that young infants can categorize their own vocalizations, which has important implications for early vocal learning.

  9. Identification of American English vowels by native Japanese speakers: Talker-and-token-based analysis

    NASA Astrophysics Data System (ADS)

    Nozawa, Takeshi; Frieda, Elaina M.; Wayland, Ratree

    2005-09-01

    Native speakers of Japanese identified American English vowels /i, I, ɛ, æ, squflg, squflg/ produced by four female native speakers of American English in /CVC/ contexts. Native speakers of American English served as the control group, and they outperformed the Japanese subjects in identifying all the English vowels in every /CVC/ context. In another experiment the Japanese subjects equated these English vowels with Japanese vowels. In general, English vowels were equated with phonetically close Japanese vowels, but significant talker effect was observed. The /i/ tokens equated with the Japanese long high front vowel /ii/ were much more correctly identified as /i/ than those equated with the Japanese short high front vowel /i/. These tokens were more often misidentified as /I/. The /squflg/ and /squflg/ tokens were predominantly equated with the Japanese low vowel /a/. The percent-correct identification of /squflg/ and /squflg/ was low in most of the /CVC/ contexts, and these two vowels were often misidentified as each other, and the Japanese subjects' latency before they decided what vowel they had heard was longer when /squflg/ or /squflg/ tokens were presented. The Japanese subjects do not seem to have salient cues to differentiate /squflg/ and /squflg/.

  10. Liquid Rocket Booster (LRB) for the Space Transportation System (STS) systems study. Appendix B: Liquid rocket booster acoustic and thermal environments

    NASA Technical Reports Server (NTRS)

    1989-01-01

    The ascent thermal environment and propulsion acoustic sources for the Martin-Marietta Corporation designed Liquid Rocket Boosters (LRB) to be used with the Space Shuttle Orbiter and External Tank are described. Two designs were proposed: one using a pump-fed propulsion system and the other using a pressure-fed propulsion system. Both designs use LOX/RP-1 propellants, but differences in performance of the two propulsion systems produce significant differences in the proposed stage geometries, exhaust plumes, and resulting environments. The general characteristics of the two designs which are significant for environmental predictions are described. The methods of analysis and predictions for environments in acoustics, aerodynamic heating, and base heating (from exhaust plume effects) are also described. The acoustic section will compare the proposed exhaust plumes with the current SRB from the standpoint of acoustics and ignition overpressure. The sections on thermal environments will provide details of the LRB heating rates and indications of possible changes in the Orbiter and ET environments as a result of the change from SRBs to LRBs.

  11. AST Launch Vehicle Acoustics

    NASA Technical Reports Server (NTRS)

    Houston, Janice; Counter, D.; Giacomoni, D.

    2015-01-01

    The liftoff phase induces acoustic loading over a broad frequency range for a launch vehicle. These external acoustic environments are then used in the prediction of internal vibration responses of the vehicle and components which result in the qualification levels. Thus, predicting these liftoff acoustic (LOA) environments is critical to the design requirements of any launch vehicle. If there is a significant amount of uncertainty in the predictions or if acoustic mitigation options must be implemented, a subscale acoustic test is a feasible pre-launch test option to verify the LOA environments. The NASA Space Launch System (SLS) program initiated the Scale Model Acoustic Test (SMAT) to verify the predicted SLS LOA environments and to determine the acoustic reduction with an above deck water sound suppression system. The SMAT was conducted at Marshall Space Flight Center and the test article included a 5% scale SLS vehicle model, tower and Mobile Launcher. Acoustic and pressure data were measured by approximately 250 instruments. The SMAT liftoff acoustic results are presented, findings are discussed and a comparison is shown to the Ares I Scale Model Acoustic Test (ASMAT) results.

  12. Lingual Electromyography Related to Tongue Movements in Swedish Vowel Production.

    ERIC Educational Resources Information Center

    Hirose, Hajime; And Others

    1979-01-01

    In order to investigate the articulatory dynamics of the tongue in the production of Swedish vowels, electromyographic (EMG) and X-ray microbeam studies were performed on a native Swedish subject. The EMG signals were used to obtain average indication of the muscle activity of the tongue as a function of time. (NCR)

  13. Stimulus Variability and Perceptual Learning of Nonnative Vowel Categories

    ERIC Educational Resources Information Center

    Brosseau-Lapre, Francoise; Rvachew, Susan; Clayards, Meghan; Dickson, Daniel

    2013-01-01

    English-speakers' learning of a French vowel contrast (/schwa/-/slashed o/) was examined under six different stimulus conditions in which contrastive and noncontrastive stimulus dimensions were varied orthogonally to each other. The distribution of contrastive cues was varied across training conditions to create single prototype, variable far…

  14. Fricatives, Affricates, and Vowels in Croatian Children with Cochlear Implants

    ERIC Educational Resources Information Center

    Mildner, Vesna; Liker, Marko

    2008-01-01

    The aim of the research was to analyse the speech of children with cochlear implants over approximately a 46-month period, and compare it with the speech of hearing controls. It focused on three categories of sounds in Croatian: vowels (F1 and F2 of /i/, /e/, /a/, /o/ and /u/), fricatives /s/ and /[esh]/ (spectral differences expressed in terms of…

  15. Does Size Matter? Subsegmental Cues to Vowel Mispronunciation Detection

    ERIC Educational Resources Information Center

    Mani, Nivedita; Plunkett, Kim

    2011-01-01

    Children look longer at a familiar object when presented with either correct pronunciations or small mispronunciations of consonants in the object's label, but not following larger mispronunciations. The current article examines whether children display a similar graded sensitivity to different degrees of mispronunciations of the vowels in…

  16. Nasal Vowels as Two Segments: Evidence from Borrowings.

    ERIC Educational Resources Information Center

    Paradis, Carole; Prunet, Jean-Francois

    2000-01-01

    Demonstrates that the substitution of a foreign segment in the borrowing of a database, which includes 14,350 segmental malformations from French and English loanwords in eight distinct languages, involves its replacement by a single native placement. This tendency is without exception, other than in cases where nasal vowels are concerned.…

  17. Vowel Harmony in Palestinian Arabic: A Metrical Perspective.

    ERIC Educational Resources Information Center

    Abu-Salim, I. M.

    1987-01-01

    The autosegmental rule of vowel harmony (VH) in Palestinian Arabic is shown to be constrained simultaneously by metrical and segmental boundaries. The indicative prefix bi- is no longer an exception to VH if a structure is assumed that disallows the prefix from sharing a foot with the stem, consequently blocking VH. (Author/LMO)

  18. Comparison of Nasal Acceleration and Nasalance across Vowels

    ERIC Educational Resources Information Center

    Thorp, Elias B.; Virnik, Boris T.; Stepp, Cara E.

    2013-01-01

    Purpose: The purpose of this study was to determine the performance of normalized nasal acceleration (NNA) relative to nasalance as estimates of nasalized versus nonnasalized vowel and sentence productions. Method: Participants were 18 healthy speakers of American English. NNA was measured using a custom sensor, and nasalance was measured using…

  19. Identification of Simple and Compound Vowels by First Graders.

    ERIC Educational Resources Information Center

    Wright, Ouida T.

    This instrument was designed to determine whether by structuring and sequencing monosyllabic English words in two different patterns--administered with the same control procedures--first-grade children would be aided in detecting, identifying, and discriminating among single vowels and their combined forms; in associating them with their specific…

  20. Vowel Epenthesis and Segment Identity in Korean Learners of English

    ERIC Educational Resources Information Center

    de Jong, Kenneth; Park, Hanyong

    2012-01-01

    Recent literature has sought to understand the presence of epenthetic vowels after the productions of postvocalic word-final consonants by second language (L2) learners whose first languages (L1s) restrict the presence of obstruents in coda position. Previous models include those in which epenthesis is seen as a strategy to mitigate the effects of…

  1. Vowel Harmony: A Variable Rule in Brazilian Portuguese.

    ERIC Educational Resources Information Center

    Bisol, Leda

    1989-01-01

    Examines vowel harmony in the "Gaucho dialect" of the Brazilian state of Rio Grande do Sul. Informants from four areas of the state were studied: the capital city (Porto Alegre), the border region with Uruguay, and two areas of the interior populated by descendants of nineteenth-century immigrants from Europe, mainly Germans and Italians. (VWL)

  2. Vowel Production in the Speech of Western Armenian Heritage Speakers

    ERIC Educational Resources Information Center

    Godson, Linda

    2004-01-01

    This study investigates whether the age at which English becomes dominant for Western Armenian bilinguals in the United States affects their vowel production in Western Armenian. Participating in the study were ten Western-Armenian bilinguals who learned English before age 8, ten bilinguals who did not learn English until adulthood, and one…

  3. Effect of Audio vs. Video on Aural Discrimination of Vowels

    ERIC Educational Resources Information Center

    McCrocklin, Shannon

    2012-01-01

    Despite the growing use of media in the classroom, the effects of using of audio versus video in pronunciation teaching has been largely ignored. To analyze the impact of the use of audio or video training on aural discrimination of vowels, 61 participants (all students at a large American university) took a pre-test followed by two training…

  4. Hemispheric Differences in the Effects of Context on Vowel Perception

    ERIC Educational Resources Information Center

    Sjerps, Matthias J.; Mitterer, Holger; McQueen, James M.

    2012-01-01

    Listeners perceive speech sounds relative to context. Contextual influences might differ over hemispheres if different types of auditory processing are lateralized. Hemispheric differences in contextual influences on vowel perception were investigated by presenting speech targets and both speech and non-speech contexts to listeners' right or left…

  5. Proficient bilinguals require more information for vowel identification than monolinguals

    NASA Astrophysics Data System (ADS)

    Lopez, Alexandra S.; Rogers, Catherine L.

    2001-05-01

    Even proficient bilinguals have been shown to experience more difficulty understanding speech in noise than monolinguals. One potential explanation is that bilinguals require more information than monolinguals for phoneme identification. We tested this hypothesis by presenting gated, silent-center vowels to two groups of listeners: (1) monolingual American English speakers and (2) proficient Spanish-English bilinguals, who spoke unaccented or mildly accented English. To create the stimuli, two American English speakers were recorded as they read the following items: ``beeb, bibb, babe, bebb, babb,'' and ``bob.'' Duration-preserved silent-center versions of three tokens of each item were created by retaining varying amounts of the CV and VC transitions (10, 20, 30, or 40 ms) and attenuating the remainder of the vowel center to silence. Duration-neutral versions of silent-center tokens were created by lengthening or shortening the silent portion to match the tokens vowel duration to the average for all the tokens. Listeners identified the unedited (full vowel), duration-preserved, and duration neutral silent-center tokens in a six-alternative forced-choice task. The two groups of listeners identified the unedited tokens with similar accuracy. In the silent-center conditions, however, the bilinguals identified the stimuli less accurately than the monolinguals. [Work supported in part by NIH-NIDCD Grant No. 1R03DC005561-01A1.

  6. Vowel Category Formation in Korean-English Bilingual Children

    ERIC Educational Resources Information Center

    Lee, Sue Ann S.; Iverson, Gregory K.

    2012-01-01

    Purpose: A previous investigation (Lee & Iverson, 2012) found that English and Korean stop categories were fully distinguished by Korean-English bilingual children at 10 years of age but not at 5 years of age. The present study examined vowels produced by Korean-English bilingual children of these same ages to determine whether and when bilinguals…

  7. Comparing Identification of Standardized and Regionally Valid Vowels

    ERIC Educational Resources Information Center

    Wright, Richard; Souza, Pamela

    2012-01-01

    Purpose: In perception studies, it is common to use vowel stimuli from standardized recordings or synthetic stimuli created using values from well-known published research. Although the use of standardized stimuli is convenient, unconsidered dialect and regional accent differences may introduce confounding effects. The goal of this study was to…

  8. Consonants and Vowels: Different Roles in Early Language Acquisition

    ERIC Educational Resources Information Center

    Hochmann, Jean-Remy; Benavides-Varela, Silvia; Nespor, Marina; Mehler, Jacques

    2011-01-01

    Language acquisition involves both acquiring a set of words (i.e. the lexicon) and learning the rules that combine them to form sentences (i.e. syntax). Here, we show that consonants are mainly involved in word processing, whereas vowels are favored for extracting and generalizing structural relations. We demonstrate that such a division of labor…

  9. Vowels in Early Words: An Event-Related Potential Study

    ERIC Educational Resources Information Center

    Mani, Nivedita; Mills, Debra L.; Plunkett, Kim

    2012-01-01

    Previous behavioural research suggests that infants possess phonologically detailed representations of the vowels and consonants in familiar words. These tasks examine infants' sensitivity to mispronunciations of a target label in the presence of a target and distracter image. Sensitivity to the mispronunciation may, therefore, be contaminated by…

  10. Competing Triggers: Transparency and Opacity in Vowel Harmony

    ERIC Educational Resources Information Center

    Kimper, Wendell A.

    2011-01-01

    This dissertation takes up the issue of transparency and opacity in vowel harmony--that is, when a segment is unable to undergo a harmony process, will it be skipped over by harmony (transparent) or will it prevent harmony from propagating further (opaque)? I argue that the choice between transparency and opacity is best understood as a…

  11. Finding Words in a Language that Allows Words without Vowels

    ERIC Educational Resources Information Center

    El Aissati, Abder; McQueen, James M.; Cutler, Anne

    2012-01-01

    Across many languages from unrelated families, spoken-word recognition is subject to a constraint whereby potential word candidates must contain a vowel. This constraint minimizes competition from embedded words (e.g., in English, disfavoring "win" in "twin" because "t" cannot be a word). However, the constraint would be counter-productive in…

  12. Auditory Spectral Integration in the Perception of Static Vowels

    ERIC Educational Resources Information Center

    Fox, Robert Allen; Jacewicz, Ewa; Chang, Chiung-Yun

    2011-01-01

    Purpose: To evaluate potential contributions of broadband spectral integration in the perception of static vowels. Specifically, can the auditory system infer formant frequency information from changes in the intensity weighting across harmonics when the formant itself is missing? Does this type of integration produce the same results in the lower…

  13. Influence of lips on the production of vowels based on finite element simulations and experiments.

    PubMed

    Arnela, Marc; Blandin, Rémi; Dabbaghchian, Saeed; Guasch, Oriol; Alías, Francesc; Pelorson, Xavier; Van Hirtum, Annemie; Engwall, Olov

    2016-05-01

    Three-dimensional (3-D) numerical approaches for voice production are currently being investigated and developed. Radiation losses produced when sound waves emanate from the mouth aperture are one of the key aspects to be modeled. When doing so, the lips are usually removed from the vocal tract geometry in order to impose a radiation impedance on a closed cross-section, which speeds up the numerical simulations compared to free-field radiation solutions. However, lips may play a significant role. In this work, the lips' effects on vowel sounds are investigated by using 3-D vocal tract geometries generated from magnetic resonance imaging. To this aim, two configurations for the vocal tract exit are considered: with lips and without lips. The acoustic behavior of each is analyzed and compared by means of time-domain finite element simulations that allow free-field wave propagation and experiments performed using 3-D-printed mechanical replicas. The results show that the lips should be included in order to correctly model vocal tract acoustics not only at high frequencies, as commonly accepted, but also in the low frequency range below 4 kHz, where plane wave propagation occurs. PMID:27250177

  14. Effects of speaking condition and hearing status on vowel production in postlingually deaf adults with cochlear implant

    NASA Astrophysics Data System (ADS)

    Menard, Lucie; Denny, Margaret; Lane, Harlan; Matthies, Melanie L.; Perkell, Joseph S.; Stockmann, Ellen; Vick, Jennell; Zandipour, Majid; Balkany, Thomas; Polack, Marek; Tiede, Mark K.

    2005-09-01

    This study investigates the effects of speaking condition and hearing status on vowel production by postlingually deafened adults. Thirteen cochlear implant users produced repetitions of nine American English vowels in three time samples (prior to implantation, one month, and one year after implantation); in three speaking conditions (clear, normal, and fast); and in two feedback conditions after implantation (implant processor turned on and off). Ten speakers with normal hearing were also recorded. Results show that vowel contrast, in the F1 versus F2 space, in mels, does not differ from the pre-implant stage to the 1-month stage. This finding indicates that shortly after implantation, speakers had not had enough experience with hearing from the implant to adequately retune their auditory feedback systems and use auditory feedback to improve feedforward commands. After 1 year of implant use, contrast distances had increased in both feedback conditions (processor on and off), indicating improvement in feedforward commands for phoneme production. Furthermore, after long-term auditory deprivation, speakers were producing differences in contrast between clear and fast conditions in the range of those found for normal-hearing speakers, leading to the inference that maintenance of these distinctions is not affected by hearing status. [Research supported by NIDCD.

  15. Capabilities, Design, Construction and Commissioning of New Vibration, Acoustic, and Electromagnetic Capabilities Added to the World's Largest Thermal Vacuum Chamber at NASA's Space Power Facility

    NASA Technical Reports Server (NTRS)

    Motil, Susan M.; Ludwiczak, Damian R.; Carek, Gerald A.; Sorge, Richard N.; Free, James M.; Cikanek, Harry A., III

    2011-01-01

    NASA s human space exploration plans developed under the Exploration System Architecture Studies in 2005 included a Crew Exploration Vehicle launched on an Ares I launch vehicle. The mass of the Crew Exploration Vehicle and trajectory of the Ares I coupled with the need to be able to abort across a large percentage of the trajectory generated unprecedented testing requirements. A future lunar lander added to projected test requirements. In 2006, the basic test plan for Orion was developed. It included several types of environment tests typical of spacecraft development programs. These included thermal-vacuum, electromagnetic interference, mechanical vibration, and acoustic tests. Because of the size of the vehicle and unprecedented acoustics, NASA conducted an extensive assessment of options for testing, and as result, chose to augment the Space Power Facility at NASA Plum Brook Station, of the John H. Glenn Research Center to provide the needed test capabilities. The augmentation included designing and building the World s highest mass capable vibration table, the highest power large acoustic chamber, and adaptation of the existing World s largest thermal vacuum chamber as a reverberant electromagnetic interference test chamber. These augmentations were accomplished from 2007 through early 2011. Acceptance testing began in Spring 2011 and will be completed in the Fall of 2011. This paper provides an overview of the capabilities, design, construction and acceptance of this extraordinary facility.

  16. The effect of vibrato on the recognition of masked vowels.

    PubMed

    Demany, L; Semal, C

    1990-11-01

    Five experiments on the identifiability of synthetic vowels masked by wideband sounds are reported. In each experiment, identification thresholds (signal/masker ratios, in decibels) were measured for two versions of four vowels: a vibrated version, in which FO varied sinusoidally around 100 Hz; and a steady version, in which F0 was fixed at 100 Hz. The first three experiments were performed on naive subjects. Experiment 1 showed that for maskers consisting of bursts of pink noise, vibrato had no effect on thresholds. In Experiment 2, where the maskers were periodic pulse trains with an F0 randomly varied between 120 and 140 Hz from trial to trial, vibrato slightly improved thresholds when the sound pressure level of the maskers was 40 dB, but had no effect for 65-dB maskers. In Experiment 3, vibrated rather than steady pulse trains were used as maskers; when these maskers were at 40 dB, the vibrated versions of the vowels were slightly less identifiable than their steady versions; but, as in Experiment 2, vibrato had no effect when the maskers were at 65 dB. Experiment 4 showed that the unmasking effect of vibrato found in Experiment 2 disappeared in subjects trained in the identification task. Finally, Experiment 5 indicated that in trained listeners, vibrato had no influence on identification performance even when the maskers and the vowels had synchronous onsets and offsets. We conclude that vibrating a vowel masked by a wideband sound can affect its identification threshold, but only for tonal maskers and in untrained listeners. This effect of vibrato should probably be considered as a Gestalt phenomenon originating from central auditory mechanisms. PMID:2247326

  17. Modeling of aerodynamic interaction between vocal folds and vocal tract during production of a vowel-voiceless plosive-vowel sequence.

    PubMed

    Delebecque, Louis; Pelorson, Xavier; Beautemps, Denis

    2016-01-01

    The context of this study is the physical modeling of speech production. The objective is, by using a mechanical replica of the vocal tract, to test quantitatively an aerodynamic model of the interaction between the vocal folds and the vocal tract during the production of a vowel-voiceless plosive-vowel sequence. The first step is to realize acoustic and aerodynamic measurements on a speaker during the production of an /apa/ sequence. The aperture and width of the lips are also derived from a high-speed video recording of the subject's face. Theoretical models to describe the flow through the lips and the effect of an expansion of the supraglottal cavity are proposed and validated by comparison with measurements made using a self-oscillating replica of the phonatory system. Finally, using these models, numerical simulations of an /apa/ sequence are performed using the measured lip parameters as the only time-varying input parameters. The results of these simulations suggest that the realization of an occlusion of the vocal tract produces a passive increase in glottal area associated with a voice offset and that the expansion of the supraglottal cavity is responsible for the extension of the phonation up to 40 ms after closure of the lips.

  18. The Acoustic and Perceptual Correlates of Emphasis in Urban Jordanian Arabic

    ERIC Educational Resources Information Center

    Al-Masri, Mohammad

    2009-01-01

    Acoustic and perceptual correlates of emphasis, a secondary articulation in the posterior vocal tract, in Urban Jordanian Arabic were studied. CVC monosyllables and CV.CVC bisyllables with emphatic and plain target consonants in word-initial, word-medial and word-final positions were examined. Spectral measurements on the target vowels at vowel…

  19. Acoustic Cues to Perception of Word Stress by English, Mandarin, and Russian Speakers

    ERIC Educational Resources Information Center

    Chrabaszcz, Anna; Winn, Matthew; Lin, Candise Y.; Idsardi, William J.

    2014-01-01

    Purpose: This study investigated how listeners' native language affects their weighting of acoustic cues (such as vowel quality, pitch, duration, and intensity) in the perception of contrastive word stress. Method: Native speakers (N = 45) of typologically diverse languages (English, Russian, and Mandarin) performed a stress identification…

  20. Articulatory-to-Acoustic Relations in Talkers with Dysarthria: A First Analysis

    ERIC Educational Resources Information Center

    Mefferd, Antje

    2015-01-01

    Purpose: The primary purpose of this study was to determine the strength of interspeaker and intraspeaker articulatory-to-acoustic relations of vowel contrast produced by talkers with dysarthria and controls. Methods: Six talkers with amyotrophic lateral sclerosis (ALS), six talkers with Parkinson's disease (PD), and 12 controls repeated a…

  1. Toward the Development of an Objective Index of Dysphonia Severity: A Four-Factor Acoustic Model

    ERIC Educational Resources Information Center

    Awan, Shaheen N.; Roy, Nelson

    2006-01-01

    During assessment and management of individuals with voice disorders, clinicians routinely attempt to describe or quantify the severity of a patient's dysphonia. This investigation used acoustic measures derived from sustained vowel samples to predict dysphonia severity (as determined by auditory-perceptual ratings), for a diverse set of voice…

  2. The Time Course for Processing Vowels and Lexical Tones: Reading Aloud Thai Words.

    PubMed

    Davis, Chris; Schoknecht, Colin; Kim, Jeesun; Burnham, Denis

    2016-06-01

    Three naming aloud experiments and a lexical decision (LD) experiment used masked priming to index the processing of written Thai vowels and tones. Thai allows for manipulation of the mapping between orthography and phonology not possible in other orthographies, for example, the use of consonants, vowels and tone markers in both horizontal and vertical orthographic positions (HOPs and VOPs). Experiment I showed that changing a vowel between prime and target slowed down target naming but changing a tone mark did not. Experiment I used an across item-design and a different number of HOPs in the way vowels and tones were specified. Experiment 2 used a within-item design and tested vowel and tone changes for both 2-HOP and 3-HOP targets separately. The 3-HOP words showed the same tone and vowel change effect as Experiment 1, whereas 2-HOP items did not. It was speculated that the 2-HOP result was due to the variable position of the vowel affecting priming. Experiment 3 employed a more stringent control over the 2-HOP vowel and tone items and found priming for the tone changes but not for vowel changes. The final experiment retested the items from Experiment 3 with the LD task and found no priming for the tone change items, indicating that the tone effect in Experiment 3 was due to processes involved in naming aloud. In all, the results supported the view that for naming a word, the development of tone information is slower than vowel information.

  3. The Time Course for Processing Vowels and Lexical Tones: Reading Aloud Thai Words.

    PubMed

    Davis, Chris; Schoknecht, Colin; Kim, Jeesun; Burnham, Denis

    2016-06-01

    Three naming aloud experiments and a lexical decision (LD) experiment used masked priming to index the processing of written Thai vowels and tones. Thai allows for manipulation of the mapping between orthography and phonology not possible in other orthographies, for example, the use of consonants, vowels and tone markers in both horizontal and vertical orthographic positions (HOPs and VOPs). Experiment I showed that changing a vowel between prime and target slowed down target naming but changing a tone mark did not. Experiment I used an across item-design and a different number of HOPs in the way vowels and tones were specified. Experiment 2 used a within-item design and tested vowel and tone changes for both 2-HOP and 3-HOP targets separately. The 3-HOP words showed the same tone and vowel change effect as Experiment 1, whereas 2-HOP items did not. It was speculated that the 2-HOP result was due to the variable position of the vowel affecting priming. Experiment 3 employed a more stringent control over the 2-HOP vowel and tone items and found priming for the tone changes but not for vowel changes. The final experiment retested the items from Experiment 3 with the LD task and found no priming for the tone change items, indicating that the tone effect in Experiment 3 was due to processes involved in naming aloud. In all, the results supported the view that for naming a word, the development of tone information is slower than vowel information. PMID:27363253

  4. Neural Representation of Concurrent Vowels in Macaque Primary Auditory Cortex123

    PubMed Central

    Micheyl, Christophe; Steinschneider, Mitchell

    2016-01-01

    Abstract Successful speech perception in real-world environments requires that the auditory system segregate competing voices that overlap in frequency and time into separate streams. Vowels are major constituents of speech and are comprised of frequencies (harmonics) that are integer multiples of a common fundamental frequency (F0). The pitch and identity of a vowel are determined by its F0 and spectral envelope (formant structure), respectively. When two spectrally overlapping vowels differing in F0 are presented concurrently, they can be readily perceived as two separate “auditory objects” with pitches at their respective F0s. A difference in pitch between two simultaneous vowels provides a powerful cue for their segregation, which in turn, facilitates their individual identification. The neural mechanisms underlying the segregation of concurrent vowels based on pitch differences are poorly understood. Here, we examine neural population responses in macaque primary auditory cortex (A1) to single and double concurrent vowels (/a/ and /i/) that differ in F0 such that they are heard as two separate auditory objects with distinct pitches. We find that neural population responses in A1 can resolve, via a rate-place code, lower harmonics of both single and double concurrent vowels. Furthermore, we show that the formant structures, and hence the identities, of single vowels can be reliably recovered from the neural representation of double concurrent vowels. We conclude that A1 contains sufficient spectral information to enable concurrent vowel segregation and identification by downstream cortical areas. PMID:27294198

  5. Sensitivity of envelope following responses to vowel polarity.

    PubMed

    Easwar, Vijayalakshmi; Beamish, Laura; Aiken, Steven; Choi, Jong Min; Scollie, Susan; Purcell, David

    2015-02-01

    Envelope following responses (EFRs) elicited by stimuli of opposite polarities are often averaged due to their insensitivity to polarity when elicited by amplitude modulated tones. A recent report illustrates that individuals exhibit varying degrees of polarity-sensitive differences in EFR amplitude when elicited by vowel stimuli (Aiken and Purcell, 2013). The aims of the current study were to evaluate the incidence and degree of polarity-sensitive differences in EFRs recorded in a large group of individuals, and to examine potential factors influencing the polarity-sensitive nature of EFRs. In Experiment I of the present study, we evaluated the incidence and degree of polarity-sensitive differences in EFR amplitude in a group of 39 participants. EFRs were elicited by opposite polarities of the vowel /ε/ in a natural /hVd/ context presented at 80 dB SPL. Nearly 30% of the participants with detectable responses (n = 24) showed a difference of greater than ∼39 nV in EFR response amplitude between the two polarities, that was unexplained by variations in noise estimates. In Experiment II, we evaluated the effect of vowel, frequency of harmonics and presence of the first harmonic (h1) on the polarity sensitivity of EFRs in 20 participants with normal hearing. For vowels /u/, /a/ and /i/, EFRs were elicited by two simultaneously presented carriers representing the first formant (resolved harmonics), and the second and higher formants (unresolved harmonics). Individual but simultaneous EFRs were elicited by the formant carriers by separating the fundamental frequency in the two carriers by 8 Hz. Vowels were presented as part of a naturally produced, but modified sequence /susaʃi/, at an overall level of 65 dB SPL. To evaluate the effect of h1 on polarity sensitivity of EFRs, EFRs were elicited by the same vowels without h1 in an identical sequence. A repeated measures analysis of variance indicated a significant effect of polarity on EFR amplitudes for the

  6. Vowel production, speech-motor control, and phonological encoding in people who are lesbian, bisexual, or gay, and people who are not

    NASA Astrophysics Data System (ADS)

    Munson, Benjamin; Deboe, Nancy

    2003-10-01

    A recent study (Pierrehumbert, Bent, Munson, and Bailey, submitted) found differences in vowel production between people who are lesbian, bisexual, or gay (LBG) and people who are not. The specific differences (more fronted /u/ and /a/ in the non-LB women; an overall more-contracted vowel space in the non-gay men) were not amenable to an interpretation based on simple group differences in vocal-tract geometry. Rather, they suggested that differences were either due to group differences in some other skill, such as motor control or phonological encoding, or learned. This paper expands on this research by examining vowel production, speech-motor control (measured by diadochokinetic rates), and phonological encoding (measured by error rates in a tongue-twister task) in people who are LBG and people who are not. Analyses focus on whether the findings of Pierrehumbert et al. (submitted) are replicable, and whether group differences in vowel production are related to group differences in speech-motor control or phonological encoding. To date, 20 LB women, 20 non-LB women, 7 gay men, and 7 non-gay men have participated. Preliminary analyses suggest that there are no group differences in speech motor control or phonological encoding, suggesting that the earlier findings of Pierrehumbert et al. reflected learned behaviors.

  7. Acoustic-Levitation Chamber

    NASA Technical Reports Server (NTRS)

    Barmatz, M. B.; Granett, D.; Lee, M. C.

    1984-01-01

    Uncontaminated environments for highly-pure material processing provided within completely sealed levitation chamber that suspends particles by acoustic excitation. Technique ideally suited for material processing in low gravity environment of space.

  8. Acoustic mode coupling induced by shallow water nonlinear internal waves: sensitivity to environmental conditions and space-time scales of internal waves.

    PubMed

    Colosi, John A

    2008-09-01

    While many results have been intuited from numerical simulation studies, the precise connections between shallow-water acoustic variability and the space-time scales of nonlinear internal waves (NLIWs) as well as the background environmental conditions have not been clearly established analytically. Two-dimensional coupled mode propagation through NLIWs is examined using a perturbation series solution in which each order n is associated with nth-order multiple scattering. Importantly, the perturbation solution gives resonance conditions that pick out specific NLIW scales that cause coupling, and seabed attenuation is demonstrated to broaden these resonances, fundamentally changing the coupling behavior at low frequency. Sound-speed inhomogeneities caused by internal solitary waves (ISWs) are primarily considered and the dependence of mode coupling on ISW amplitude, range width, depth structure, location relative to the source, and packet characteristics are delineated as a function of acoustic frequency. In addition, it is seen that significant energy transfer to modes with initially low or zero energy involves at least a second order scattering process. Under moderate scattering conditions, comparisons of first order, single scattering theoretical predictions to direct numerical simulation demonstrate the accuracy of the approach for acoustic frequencies upto 400 Hz and for single as well as multiple ISW wave packets.

  9. Automated Classification of Vowel Category and Speaker Type in the High-Frequency Spectrum.

    PubMed

    Donai, Jeremy J; Motiian, Saeid; Doretto, Gianfranco

    2016-04-20

    The high-frequency region of vowel signals (above the third formant or F3) has received little research attention. Recent evidence, however, has documented the perceptual utility of high-frequency information in the speech signal above the traditional frequency bandwidth known to contain important cues for speech and speaker recognition. The purpose of this study was to determine if high-pass filtered vowels could be separated by vowel category and speaker type in a supervised learning framework. Mel frequency cepstral coefficients (MFCCs) were extracted from productions of six vowel categories produced by two male, two female, and two child speakers. Results revealed that the filtered vowels were well separated by vowel category and speaker type using MFCCs from the high-frequency spectrum. This demonstrates the presence of useful information for automated classification from the high-frequency region and is the first study to report findings of this nature in a supervised learning framework. PMID:27588160

  10. Automated Classification of Vowel Category and Speaker Type in the High-Frequency Spectrum.

    PubMed

    Donai, Jeremy J; Motiian, Saeid; Doretto, Gianfranco

    2016-04-20

    The high-frequency region of vowel signals (above the third formant or F3) has received little research attention. Recent evidence, however, has documented the perceptual utility of high-frequency information in the speech signal above the traditional frequency bandwidth known to contain important cues for speech and speaker recognition. The purpose of this study was to determine if high-pass filtered vowels could be separated by vowel category and speaker type in a supervised learning framework. Mel frequency cepstral coefficients (MFCCs) were extracted from productions of six vowel categories produced by two male, two female, and two child speakers. Results revealed that the filtered vowels were well separated by vowel category and speaker type using MFCCs from the high-frequency spectrum. This demonstrates the presence of useful information for automated classification from the high-frequency region and is the first study to report findings of this nature in a supervised learning framework.

  11. Automated Classification of Vowel Category and Speaker Type in the High-Frequency Spectrum

    PubMed Central

    Donai, Jeremy J.; Motiian, Saeid; Doretto, Gianfranco

    2016-01-01

    The high-frequency region of vowel signals (above the third formant or F3) has received little research attention. Recent evidence, however, has documented the perceptual utility of high-frequency information in the speech signal above the traditional frequency bandwidth known to contain important cues for speech and speaker recognition. The purpose of this study was to determine if high-pass filtered vowels could be separated by vowel category and speaker type in a supervised learning framework. Mel frequency cepstral coefficients (MFCCs) were extracted from productions of six vowel categories produced by two male, two female, and two child speakers. Results revealed that the filtered vowels were well separated by vowel category and speaker type using MFCCs from the high-frequency spectrum. This demonstrates the presence of useful information for automated classification from the high-frequency region and is the first study to report findings of this nature in a supervised learning framework. PMID:27588160

  12. Constraints of Tones, Vowels and Consonants on Lexical Selection in Mandarin Chinese.

    PubMed

    Wiener, Seth; Turnbull, Rory

    2016-03-01

    Previous studies have shown that when speakers of European languages are asked to turn nonwords into words by altering either a vowel or consonant, they tend to treat vowels as more mutable than consonants. These results inspired the universal vowel mutability hypothesis: listeners learn to cope with vowel variability because vowel information constrains lexical selection less tightly and allows for more potential candidates than does consonant information. The present study extends the word reconstruction paradigm to Mandarin Chinese--a Sino-Tibetan language, which makes use of lexically contrastive tone. Native speakers listened to word-like nonwords (e.g., su3) and were asked to change them into words by manipulating a single consonant (e.g., tu3), vowel (e.g., si3), or tone (e.g., su4). Additionally, items were presented in a fourth condition in which participants could change any part. The participants' reaction times and responses were recorded. Results revealed that participants responded faster and more accurately in both the free response and the tonal change conditions. Unlike previous reconstruction studies on European languages, where vowels were changed faster and more often than consonants, these results demonstrate that, in Mandarin, changes to vowels and consonants were both overshadowed by changes to tone, which was the preferred modification to the stimulus nonwords, while changes to vowels were the slowest and least accurate. Our findings show that the universal vowel mutability hypothesis is not consistent with a tonal language, that Mandarin tonal information is lower-priority than consonants and vowels and that vowel information most tightly constrains Mandarin lexical access. PMID:27089806

  13. Vowel devoicing and the perception of spoken Japanese words.

    PubMed

    Cutler, Anne; Otake, Takashi; McQueen, James M

    2009-03-01

    Three experiments, in which Japanese listeners detected Japanese words embedded in nonsense sequences, examined the perceptual consequences of vowel devoicing in that language. Since vowelless sequences disrupt speech segmentation [Norris et al. (1997). Cognit. Psychol. 34, 191-243], devoicing is potentially problematic for perception. Words in initial position in nonsense sequences were detected more easily when followed by a sequence containing a vowel than by a vowelless segment (with or without further context), and vowelless segments that were potential devoicing environments were no easier than those not allowing devoicing. Thus asa, "morning," was easier in asau or asazu than in all of asap, asapdo, asaf, or asafte, despite the fact that the /f/ in the latter two is a possible realization of fu, with devoiced [u]. Japanese listeners thus do not treat devoicing contexts as if they always contain vowels. Words in final position in nonsense sequences, however, produced a different pattern: here, preceding vowelless contexts allowing devoicing impeded word detection less strongly (so, sake was detected less accurately, but not less rapidly, in nyaksake-possibly arising from nyakusake-than in nyagusake). This is consistent with listeners treating consonant sequences as potential realizations of parts of existing lexical candidates wherever possible.

  14. Perception of speaker size and sex of vowel sounds

    NASA Astrophysics Data System (ADS)

    Smith, David R. R.; Patterson, Roy D.

    2005-04-01

    Glottal-pulse rate (GPR) and vocal-tract length (VTL) are both related to speaker size and sex-however, it is unclear how they interact to determine our perception of speaker size and sex. Experiments were designed to measure the relative contribution of GPR and VTL to judgements of speaker size and sex. Vowels were scaled to represent people with different GPRs and VTLs, including many well beyond the normal population values. In a single interval, two response rating paradigm, listeners judged the size (using a 7-point scale) and sex/age of the speaker (man, woman, boy, or girl) of these scaled vowels. Results from the size-rating experiments show that VTL has a much greater influence upon judgements of speaker size than GPR. Results from the sex-categorization experiments show that judgements of speaker sex are influenced about equally by GPR and VTL for vowels with normal GPR and VTL values. For abnormal combinations of GPR and VTL, where low GPRs are combined with short VTLs, VTL has more influence than GPR in sex judgements. [Work supported by the UK MRC (G9901257) and the German Volkswagen Foundation (VWF 1/79 783).

  15. Speaker-Sex Discrimination for Voiced and Whispered Vowels at Short Durations

    PubMed Central

    2016-01-01

    Whispered vowels, produced with no vocal fold vibration, lack the periodic temporal fine structure which in voiced vowels underlies the perceptual attribute of pitch (a salient auditory cue to speaker sex). Voiced vowels possess no temporal fine structure at very short durations (below two glottal cycles). The prediction was that speaker-sex discrimination performance for whispered and voiced vowels would be similar for very short durations but, as stimulus duration increases, voiced vowel performance would improve relative to whispered vowel performance as pitch information becomes available. This pattern of results was shown for women’s but not for men’s voices. A whispered vowel needs to have a duration three times longer than a voiced vowel before listeners can reliably tell whether it’s spoken by a man or woman (∼30 ms vs. ∼10 ms). Listeners were half as sensitive to information about speaker-sex when it is carried by whispered compared with voiced vowels. PMID:27757218

  16. Acoustic correlates of Georgian stops

    NASA Astrophysics Data System (ADS)

    Wysocki, Tamra M.

    2003-04-01

    This paper presents results from an acoustic analysis of Georgian stops. Georgian is a South Caucasian language and has a three-way opposition among voiced, voiceless aspirated, and ejective stops for three places of articulation: bilabial, dental, and velar. Tokens consist of initial and medial stops in isolated words produced by two male and two female native Georgian speakers. Closure duration, VOT, and burst amplitude were measured using waveforms, spectrograms, and FFTs. The voice quality of adjacent vowels was examined for possible consonantal effects (e.g., creaky voicing adjacent to ejectives). Additional observations included noise characteristics during closure and following oral release. Results show much variation in the realizations of stops both within and across speakers. While a general trend exists for VOT and burst amplitude to differentiate initial voiced and voiceless stops, values of voiceless aspirated and ejective stops overlap. In intervocalic stops, VOT and burst amplitude are more variable. Closure duration remains fairly stable across the three stop categories. Noise quality following oral release and, to some extent, voice quality of a following vowel more consistently distinguish the three stop types in both positions than do the quantitative measures. These characteristics are further discussed in relation to similar cross-linguistic studies.

  17. Advanced Distributed Measurements and Data Processing at the Vibro-Acoustic Test Facility, GRC Space Power Facility, Sandusky, Ohio - an Architecture and an Example

    NASA Technical Reports Server (NTRS)

    Hill, Gerald M.; Evans, Richard K.

    2009-01-01

    A large-scale, distributed, high-speed data acquisition system (HSDAS) is currently being installed at the Space Power Facility (SPF) at NASA Glenn Research Center s Plum Brook Station in Sandusky, OH. This installation is being done as part of a facility construction project to add Vibro-acoustic Test Capabilities (VTC) to the current thermal-vacuum testing capability of SPF in support of the Orion Project s requirement for Space Environments Testing (SET). The HSDAS architecture is a modular design, which utilizes fully-remotely managed components, enables the system to support multiple test locations with a wide-range of measurement types and a very large system channel count. The architecture of the system is presented along with details on system scalability and measurement verification. In addition, the ability of the system to automate many of its processes such as measurement verification and measurement system analysis is also discussed.

  18. Encoding of vowel-like sounds in the auditory nerve: Model predictions of discrimination performance

    NASA Astrophysics Data System (ADS)

    Tan, Qing; Carney, Laurel H.

    2005-03-01

    The sensitivity of listeners to changes in the center frequency of vowel-like harmonic complexes as a function of the center frequency of the complex cannot be explained by changes in the level of the stimulus [Lyzenga and Horst, J. Acoust. Soc. Am. 98, 1943-1955 (1995)]. Rather, a complex pattern of sensitivity is seen; for a spectrum with a triangular envelope, the greatest sensitivity occurs when the center frequency falls between harmonics, whereas for a spectrum with a trapezoidal envelope, greatest sensitivity occurs when the center frequency is aligned with a harmonic. In this study, the thresholds of a population model of auditory-nerve (AN) fibers were quantitatively compared to these trends in psychophysical thresholds. Single-fiber and population model responses were evaluated in terms of both average discharge rate and the combination of rate and timing information. Results indicate that phase-locked responses of AN fibers encode phase transitions associated with minima in these amplitude-modulated stimuli. The temporal response properties of a single AN fiber, tuned to a frequency slightly above the center frequency of the harmonic complex, were able to explain the trends in thresholds for both triangular- and trapezoidal-shaped spectra. .

  19. Pre-attentive sensitivity to vowel duration reveals native phonology and predicts learning of second-language sounds.

    PubMed

    Chládková, Kateřina; Escudero, Paola; Lipski, Silvia C

    2013-09-01

    In some languages (e.g. Czech), changes in vowel duration affect word meaning, while in others (e.g. Spanish) they do not. Yet for other languages (e.g. Dutch), the linguistic role of vowel duration remains unclear. To reveal whether Dutch represents vowel length in its phonology, we compared auditory pre-attentive duration processing in native and non-native vowels across Dutch, Czech, and Spanish. Dutch duration sensitivity patterned with Czech but was larger than Spanish in the native vowel, while it was smaller than Czech and Spanish in the non-native vowel. An interpretation of these findings suggests that in Dutch, duration is used phonemically but it might be relevant for the identity of certain native vowels only. Furthermore, the finding that Spanish listeners are more sensitive to duration in non-native than in native vowels indicates that a lack of duration differences in one's native language could be beneficial for second-language learning.

  20. The Role of Secondary-Stressed and Unstressed-Unreduced Syllables in Word Recognition: Acoustic and Perceptual Studies with Russian Learners of English.

    PubMed

    Banzina, Elina; Dilley, Laura C; Hewitt, Lynne E

    2016-08-01

    The importance of secondary-stressed (SS) and unstressed-unreduced (UU) syllable accuracy for spoken word recognition in English is as yet unclear. An acoustic study first investigated Russian learners' of English production of SS and UU syllables. Significant vowel quality and duration reductions in Russian-spoken SS and UU vowels were found, likely due to a transfer of native phonological features. Next, a cross-modal phonological priming technique combined with a lexical decision task assessed the effect of inaccurate SS and UU syllable productions on native American English listeners' speech processing. Inaccurate UU vowels led to significant inhibition of lexical access, while reduced SS vowels revealed less interference. The results have implications for understanding the role of SS and UU syllables for word recognition and English pronunciation instruction.

  1. Acoustic-Phonetic Differences between Infant- and Adult-Directed Speech: The Role of Stress and Utterance Position

    ERIC Educational Resources Information Center

    Wang, Yuanyuan; Seidl, Amanda; Cristia, Alejandrina

    2015-01-01

    Previous studies have shown that infant-directed speech (IDS) differs from adult-directed speech (ADS) on a variety of dimensions. The aim of the current study was to investigate whether acoustic differences between IDS and ADS in English are modulated by prosodic structure. We compared vowels across the two registers (IDS, ADS) in both stressed…

  2. Study of acoustic correlates associate with emotional speech

    NASA Astrophysics Data System (ADS)

    Yildirim, Serdar; Lee, Sungbok; Lee, Chul Min; Bulut, Murtaza; Busso, Carlos; Kazemzadeh, Ebrahim; Narayanan, Shrikanth

    2004-10-01

    This study investigates the acoustic characteristics of four different emotions expressed in speech. The aim is to obtain detailed acoustic knowledge on how a speech signal is modulated by changes from neutral to a certain emotional state. Such knowledge is necessary for automatic emotion recognition and classification and emotional speech synthesis. Speech data obtained from two semi-professional actresses are analyzed and compared. Each subject produces 211 sentences with four different emotions; neutral, sad, angry, happy. We analyze changes in temporal and acoustic parameters such as magnitude and variability of segmental duration, fundamental frequency and the first three formant frequencies as a function of emotion. Acoustic differences among the emotions are also explored with mutual information computation, multidimensional scaling and acoustic likelihood comparison with normal speech. Results indicate that speech associated with anger and happiness is characterized by longer duration, shorter interword silence, higher pitch and rms energy with wider ranges. Sadness is distinguished from other emotions by lower rms energy and longer interword silence. Interestingly, the difference in formant pattern between [happiness/anger] and [neutral/sadness] are better reflected in back vowels such as /a/(/father/) than in front vowels. Detailed results on intra- and interspeaker variability will be reported.

  3. Infant Directed Speech in Natural Interaction--Norwegian Vowel Quantity and Quality

    ERIC Educational Resources Information Center

    Englund, Kjellrun T.; Behne, Dawn M.

    2005-01-01

    An interactive face-to-face setting is used to study natural infant directed speech (IDS) compared to adult directed speech (ADS). With distinctive vowel quantity and vowel quality, Norwegian IDS was used in a natural quasi-experimental design. Six Norwegian mothers were recorded over a period of 6 months alone with their infants and in an adult…

  4. Phonology, Decoding, and Lexical Compensation in Vowel Spelling Errors Made by Children with Dyslexia

    ERIC Educational Resources Information Center

    Bernstein, Stuart E.

    2009-01-01

    A descriptive study of vowel spelling errors made by children first diagnosed with dyslexia (n = 79) revealed that phonological errors, such as "bet" for "bat", outnumbered orthographic errors, such as "bate" for "bait". These errors were more frequent in nonwords than words, suggesting that lexical context helps with vowel spelling. In a second…

  5. Effect of Syllable-Initial Voicing on Vowel Duration during Simultaneous Communication

    ERIC Educational Resources Information Center

    Metz, Dale Evan; Allen, Kristin; Kling, Therese; Maisonet, Sarah; McCullough, Rosemary; Schiavetti, Nicholas; Whitehead, Robert L.

    2006-01-01

    Vowel durations following the production of voiced and voiceless stop consonants produced during simultaneous communication (SC) were investigated by recording sign language users during SC and speech alone (SA). Under natural speaking conditions, or speaking alone (SA), vowels following voiced stop consonants are longer in duration than vowels…

  6. Shallow and Deep Orthographies in Hebrew: The Role of Vowelization in Reading Development for Unvowelized Scripts

    ERIC Educational Resources Information Center

    Schiff, Rachel

    2012-01-01

    The present study explored the speed, accuracy, and reading comprehension of vowelized versus unvowelized scripts among 126 native Hebrew speaking children in second, fourth, and sixth grades. Findings indicated that second graders read and comprehended vowelized scripts significantly more accurately and more quickly than unvowelized scripts,…

  7. Vowel Confusion Patterns in Adults during Initial 4 Years of Implant Use

    ERIC Educational Resources Information Center

    Vaalimaa, Taina T.; Sorri, Martti J.; Laitakari, Jaakko; Sivonen, Ville; Muhli, Arto

    2011-01-01

    This study investigated adult cochlear implant users' (n == 39) vowel recognition and confusions by an open-set syllable test during 4 years of implant use, in a prospective repeated-measures design. Subjects' responses were coded for phoneme errors and estimated by the generalized mixed model. Improvement in overall vowel recognition was highest…

  8. Linguistic Variation in a Border Town: Palatalization of Dental Stops and Vowel Nasalization in Rivera

    ERIC Educational Resources Information Center

    Castaneda-Molla, Rosa Maria

    2011-01-01

    This study focuses on the analysis of variation at the phonological level, specifically the variable realization of palatalization of dental stops before the high vowel /i/ and vowel nasalization in the speech of bilingual speakers of Uruguayan Portuguese (UP) in the city of Rivera, Uruguay. The data were collected in participant-observation and…

  9. Cross-Linguistic Differences in the Immediate Serial Recall of Consonants versus Vowels

    ERIC Educational Resources Information Center

    Kissling, Elizabeth M.

    2012-01-01

    The current study investigated native English and native Arabic speakers' phonological short-term memory for sequences of consonants and vowels. Phonological short-term memory was assessed in immediate serial recall tasks conducted in Arabic and English for both groups. Participants (n = 39) heard series of six consonant-vowel syllables and wrote…

  10. Effects of Short- and Long-Term Changes in Auditory Feedback on Vowel and Sibilant Contrasts

    ERIC Educational Resources Information Center

    Lane, Harlan; Matthies, Melanie L.; Guenther, Frank H.; Denny, Margaret; Perkell, Joseph S.; Stockmann, Ellen; Tiede, Mark; Vick, Jennell; Zandipour, Majid

    2007-01-01

    Purpose: To assess the effects of short- and long-term changes in auditory feedback on vowel and sibilant contrasts and to evaluate hypotheses arising from a model of speech motor planning. Method: The perception and production of vowel and sibilant contrasts were measured in 8 postlingually deafened adults prior to activation of their cochlear…

  11. Asynchronous Vowel-Pair Identification across the Adult Life Span for Monaural and Dichotic Presentations

    ERIC Educational Resources Information Center

    Fogerty, Daniel; Kewley-Port, Diane; Humes, Larry E.

    2012-01-01

    Purpose: Temporal order abilities decrease with age. Declining temporal processing abilities may influence the identification of rapid vowel sequences. Identification patterns for asynchronous vowel pairs were explored across the life span. Method: Young, middle-aged, and older listeners completed temporal order tasks for pairs of 70-ms and 40-ms…

  12. Evaluating Computational Models in Cognitive Neuropsychology: The Case from the Consonant/Vowel Distinction

    ERIC Educational Resources Information Center

    Knobel, Mark; Caramazza, Alfonso

    2007-01-01

    Caramazza et al. [Caramazza, A., Chialant, D., Capasso, R., & Miceli, G. (2000). Separable processing of consonants and vowels. "Nature," 403(6768), 428-430.] report two patients who exhibit a double dissociation between consonants and vowels in speech production. The patterning of this double dissociation cannot be explained by appealing to…

  13. Spoken word recognition in early childhood: Comparative effects of vowel, consonant and lexical tone variation.

    PubMed

    Singh, Leher; Goh, Hwee Hwee; Wewalaarachchi, Thilanga D

    2015-09-01

    The majority of the world's languages exploit consonants, vowels and lexical tones to contrast the meanings of individual words. However, the majority of experimental research on early language development focuses on consonant-vowel languages. In the present study, the role of consonants, vowels and lexical tones in emergent word knowledge are directly compared in toddlers (2.5-3.5 years) and preschoolers (4-5 years) who were bilingual native learners of a consonant-vowel-tone language (Mandarin Chinese). Using a preferential looking paradigm, participants were presented with correct pronunciations and consonantal, vowel, and tonal variations of known words. Responses to each type of variation were assessed via gaze fixations to a visual target. When their labels were correctly pronounced, visual targets were reliably identified at both age groups. However, in toddlers, there was a high degree of sensitivity to mispronunciations due to variation in lexical tones relative to those due to consonants and vowels. This pattern was reversed in preschoolers, who were more sensitive to consonant and vowel variation than to tone variation. Findings are discussed in terms of properties of tones, vowels and consonants and the respective role of each source of variation in tone languages.

  14. Shallow and deep orthographies in Hebrew: the role of vowelization in reading development for unvowelized scripts.

    PubMed

    Schiff, Rachel

    2012-12-01

    The present study explored the speed, accuracy, and reading comprehension of vowelized versus unvowelized scripts among 126 native Hebrew speaking children in second, fourth, and sixth grades. Findings indicated that second graders read and comprehended vowelized scripts significantly more accurately and more quickly than unvowelized scripts, whereas among fourth and sixth graders reading of unvowelized scripts developed to a greater degree than the reading of vowelized scripts. An analysis of the mediation effect for children's mastery of vowelized reading speed and accuracy on their mastery of unvowelized reading speed and comprehension revealed that in second grade, reading accuracy of vowelized words mediated the reading speed and comprehension of unvowelized scripts. In the fourth grade, accuracy in reading both vowelized and unvowelized words mediated the reading speed and comprehension of unvowelized scripts. By sixth grade, accuracy in reading vowelized words offered no mediating effect, either on reading speed or comprehension of unvowelized scripts. The current outcomes thus suggest that young Hebrew readers undergo a scaffolding process, where vowelization serves as the foundation for building initial reading abilities and is essential for successful and meaningful decoding of unvowelized scripts.

  15. The Prosodic Licensing of Coda Consonants in Early Speech: Interactions with Vowel Length

    ERIC Educational Resources Information Center

    Miles, Kelly; Yuen, Ivan; Cox, Felicity; Demuth, Katherine

    2016-01-01

    English has a word-minimality requirement that all open-class lexical items must contain at least two moras of structure, forming a bimoraic foot (Hayes, 1995).Thus, a word with either a long vowel, or a short vowel and a coda consonant, satisfies this requirement. This raises the question of when and how young children might learn this…

  16. The Effects of Surgical Rapid Maxillary Expansion (SRME) on Vowel Formants

    ERIC Educational Resources Information Center

    Sari, Emel; Kilic, Mehmet Akif

    2009-01-01

    The objective of this study was to investigate the effect of surgical rapid maxillary expansion (SRME) on vowel production. The subjects included 12 patients, whose speech were considered perceptually normal, that had undergone surgical RME for expansion of a narrow maxilla. They uttered the following Turkish vowels, ([a], [[epsilon

  17. An Index of Phonic Patterns by Vowel Types. AVKO "Great Idea" Reprint Series No. 622.

    ERIC Educational Resources Information Center

    McCabe, Don

    Intended for the use of teachers or diagnosticians, this booklet presents charts that list various phonic patterns, word families, or "rimes" associated with specific vowel patterns. Lists in the booklet are arranged according to the 14 basic vowel phonemes in English (including long a, long e, long i, long aw, short ah, and short u). The…

  18. Regional Dialect Variation in the Vowel Systems of Typically Developing Children

    ERIC Educational Resources Information Center

    Jacewicz, Ewa; Fox, Robert Allen; Salmons, Joseph

    2011-01-01

    Purpose: To investigate regional dialect variation in the vowel systems of typically developing 8- to 12-year-old children. Method: Thirteen vowels in isolated "h_d" words were produced by 94 children and 93 adults (males and females). All participants spoke American English and were born and raised in 1 of 3 distinct dialect regions in the United…

  19. Shallow and deep orthographies in Hebrew: the role of vowelization in reading development for unvowelized scripts.

    PubMed

    Schiff, Rachel

    2012-12-01

    The present study explored the speed, accuracy, and reading comprehension of vowelized versus unvowelized scripts among 126 native Hebrew speaking children in second, fourth, and sixth grades. Findings indicated that second graders read and comprehended vowelized scripts significantly more accurately and more quickly than unvowelized scripts, whereas among fourth and sixth graders reading of unvowelized scripts developed to a greater degree than the reading of vowelized scripts. An analysis of the mediation effect for children's mastery of vowelized reading speed and accuracy on their mastery of unvowelized reading speed and comprehension revealed that in second grade, reading accuracy of vowelized words mediated the reading speed and comprehension of unvowelized scripts. In the fourth grade, accuracy in reading both vowelized and unvowelized words mediated the reading speed and comprehension of unvowelized scripts. By sixth grade, accuracy in reading vowelized words offered no mediating effect, either on reading speed or comprehension of unvowelized scripts. The current outcomes thus suggest that young Hebrew readers undergo a scaffolding process, where vowelization serves as the foundation for building initial reading abilities and is essential for successful and meaningful decoding of unvowelized scripts. PMID:22210537

  20. Orthographic Context Sensitivity in Vowel Decoding by Portuguese Monolingual and Portuguese-English Bilingual Children

    ERIC Educational Resources Information Center

    Vale, Ana Paula

    2011-01-01

    This study examines the pronunciation of the first vowel in decoding disyllabic pseudowords derived from Portuguese words. Participants were 96 Portuguese monolinguals and 52 Portuguese-English bilinguals of equivalent Portuguese reading levels. The results indicate that sensitivity to vowel context emerges early, both in monolinguals and in…

  1. Children's Perception of Conversational and Clear American-English Vowels in Noise

    ERIC Educational Resources Information Center

    Leone, Dorothy; Levy, Erika S.

    2015-01-01

    Purpose: Much of a child's day is spent listening to speech in the presence of background noise. Although accurate vowel perception is important for listeners' accurate speech perception and comprehension, little is known about children's vowel perception in noise. "Clear speech" is a speech style frequently used by talkers in the…

  2. Perception and Production of Five English Front Vowels by College Students

    ERIC Educational Resources Information Center

    Lin, Ching-Ying

    2014-01-01

    This study was to explore whether college students could perceive and produce five English front vowels well or not. It also examined the relationship between English speaking and listening. To be more specific, the study attempted to probe which vowels that learners could be confused easily in speaking and listening. The results revealed that…

  3. Strict Cyclicity, Structure Preservation, and the Scottish Vowel-Length Rule.

    ERIC Educational Resources Information Center

    Carr, Phillip

    1992-01-01

    Reviews problems and advantages of approaches to analysis of vowel length in Standard Scottish English and Scots dialects. It is suggested that insufficient attention has been paid to operation of Scottish Vowel-Length Rule at level 1 of the lexicon, in ablaut past-tense forms, and with noun plural fricative voicing and that consideration of these…

  4. Point Vowel Duration in Children with Hearing Aids and Cochlear Implants at 4 and 5 Years of Age

    ERIC Educational Resources Information Center

    Vandam, Mark; Ide-Helvie, Dana; Moeller, Mary Pat

    2011-01-01

    This work investigates the developmental aspects of the duration of point vowels in children with normal hearing compared with those with hearing aids and cochlear implants at 4 and 5 years of age. Younger children produced longer vowels than older children, and children with hearing loss (HL) produced longer and more variable vowels than their…

  5. Bayesian State-Space Modelling of Conventional Acoustic Tracking Provides Accurate Descriptors of Home Range Behavior in a Small-Bodied Coastal Fish Species

    PubMed Central

    Alós, Josep; Palmer, Miquel; Balle, Salvador; Arlinghaus, Robert

    2016-01-01

    State-space models (SSM) are increasingly applied in studies involving biotelemetry-generated positional data because they are able to estimate movement parameters from positions that are unobserved or have been observed with non-negligible observational error. Popular telemetry systems in marine coastal fish consist of arrays of omnidirectional acoustic receivers, which generate a multivariate time-series of detection events across the tracking period. Here we report a novel Bayesian fitting of a SSM application that couples mechanistic movement properties within a home range (a specific case of random walk weighted by an Ornstein-Uhlenbeck process) with a model of observational error typical for data obtained from acoustic receiver arrays. We explored the performance and accuracy of the approach through simulation modelling and extensive sensitivity analyses of the effects of various configurations of movement properties and time-steps among positions. Model results show an accurate and unbiased estimation of the movement parameters, and in most cases the simulated movement parameters were properly retrieved. Only in extreme situations (when fast swimming speeds are combined with pooling the number of detections over long time-steps) the model produced some bias that needs to be accounted for in field applications. Our method was subsequently applied to real acoustic tracking data collected from a small marine coastal fish species, the pearly razorfish, Xyrichtys novacula. The Bayesian SSM we present here constitutes an alternative for those used to the Bayesian way of reasoning. Our Bayesian SSM can be easily adapted and generalized to any species, thereby allowing studies in freely roaming animals on the ecological and evolutionary consequences of home ranges and territory establishment, both in fishes and in other taxa. PMID:27119718

  6. Acoustical heat pumping engine

    DOEpatents

    Wheatley, John C.; Swift, Gregory W.; Migliori, Albert

    1983-08-16

    The disclosure is directed to an acoustical heat pumping engine without moving seals. A tubular housing holds a compressible fluid capable of supporting an acoustical standing wave. An acoustical driver is disposed at one end of the housing and the other end is capped. A second thermodynamic medium is disposed in the housing near to but spaced from the capped end. Heat is pumped along the second thermodynamic medium toward the capped end as a consequence both of the pressure oscillation due to the driver and imperfect thermal contact between the fluid and the second thermodynamic medium.

  7. Acoustical heat pumping engine

    DOEpatents

    Wheatley, J.C.; Swift, G.W.; Migliori, A.

    1983-08-16

    The disclosure is directed to an acoustical heat pumping engine without moving seals. A tubular housing holds a compressible fluid capable of supporting an acoustical standing wave. An acoustical driver is disposed at one end of the housing and the other end is capped. A second thermodynamic medium is disposed in the housing near to but spaced from the capped end. Heat is pumped along the second thermodynamic medium toward the capped end as a consequence both of the pressure oscillation due to the driver and imperfect thermal contact between the fluid and the second thermodynamic medium. 2 figs.

  8. Differential processing of consonants and vowels in lexical access through reading.

    PubMed

    New, Boris; Araújo, Verónica; Nazzi, Thierry

    2008-12-01

    Do consonants and vowels have the same importance during reading? Recently, it has been proposed that consonants play a more important role than vowels for language acquisition and adult speech processing. This proposal has started receiving developmental support from studies showing that infants are better at processing specific consonantal than vocalic information while learning new words. This proposal also received support from adult speech processing. In our study, we directly investigated the relative contributions of consonants and vowels to lexical access while reading by using a visual masked-priming lexical decision task. Test items were presented following four different primes: identity (e.g., for the word joli, joli), unrelated (vabu), consonant-related (jalu), and vowel-related (vobi). Priming was found for the identity and consonant-related conditions, but not for the vowel-related condition. These results establish the privileged role of consonants during lexical access while reading. PMID:19121127

  9. Consonants are More Important than Vowels in the Bouba-kiki Effect.

    PubMed

    Fort, Mathilde; Martin, Alexander; Peperkamp, Sharon

    2015-06-01

    Adult listeners systematically associate certain speech sounds with round or spiky shapes, a sound-symbolic phenomenon known as the "bouba-kiki effect." In this study, we investigate the respective influences of consonants and vowels in this phenomenon. French participants were asked to match auditorily presented pseudowords with one of two visually presented shapes, one round and one spiky. The pseudowords were created by crossing either two consonant pairs with a wide range of vowels (experiment 1 and 2) or two vowel pairs with a wide range of consonants (experiment 3). Analyses showed that consonants have a greater influence than vowels in the bouba-kiki effect. Importantly, this asymmetry cannot be due to an onset bias, as a strong consonantal influence is found both with CVCV (experiment 1) and VCV (experiment 2) stimuli. We discuss these results in terms of the differential role of consonants and vowels in speech perception. PMID:26677645

  10. Pre-attentive detection of vowel contrasts utilizes both phonetic and auditory memory representations.

    PubMed

    Winkler, I; Lehtokoski, A; Alku, P; Vainio, M; Czigler, I; Csépe, V; Aaltonen, O; Raimo, I; Alho, K; Lang, H; Iivonen, A; Näätänen, R

    1999-01-01

    Event-related brain potentials (ERP) were recorded to infrequent changes of a synthesized vowel (standard) to another vowel (deviant) in speakers of Hungarian and Finnish language, which are remotely related to each other with rather similar vowel systems. Both language groups were presented with identical stimuli. One standard-deviant pair represented an across-vowel category contrast in Hungarian, but a within-category contrast in Finnish, with the other pair having the reversed role in the two languages. Both within- and across-category contrasts elicited the mismatch negativity (MMN) ERP component in the native speakers of either language. The MMN amplitude was larger in across- than within-category contrasts in both language groups. These results suggest that the pre-attentive change-detection process generating the MMN utilized both auditory (sensory) and phonetic (categorical) representations of the test vowels. PMID:9838192

  11. Detecting Nasal Vowels in Speech Interfaces Based on Surface Electromyography

    PubMed Central

    Freitas, João; Teixeira, António; Silva, Samuel; Oliveira, Catarina; Dias, Miguel Sales

    2015-01-01

    Nasality is a very important characteristic of several languages, European Portuguese being one of them. This paper addresses the challenge of nasality detection in surface electromyography (EMG) based speech interfaces. We explore the existence of useful information about the velum movement and also assess if muscles deeper down in the face and neck region can be measured using surface electrodes, and the best electrode location to do so. The procedure we adopted uses Real-Time Magnetic Resonance Imaging (RT-MRI), collected from a set of speakers, providing a method to interpret EMG data. By ensuring compatible data recording conditions, and proper time alignment between the EMG and the RT-MRI data, we are able to accurately estimate the time when the velum moves and the type of movement when a nasal vowel occurs. The combination of these two sources revealed interesting and distinct characteristics in the EMG signal when a nasal vowel is uttered, which motivated a classification experiment. Overall results of this experiment provide evidence that it is possible to detect velum movement using sensors positioned below the ear, between mastoid process and the mandible, in the upper neck region. In a frame-based classification scenario, error rates as low as 32.5% for all speakers and 23.4% for the best speaker have been achieved, for nasal vowel detection. This outcome stands as an encouraging result, fostering the grounds for deeper exploration of the proposed approach as a promising route to the development of an EMG-based speech interface for languages with strong nasal characteristics. PMID:26069968

  12. Detecting Nasal Vowels in Speech Interfaces Based on Surface Electromyography.

    PubMed

    Freitas, João; Teixeira, António; Silva, Samuel; Oliveira, Catarina; Dias, Miguel Sales

    2015-01-01

    Nasality is a very important characteristic of several languages, European Portuguese being one of them. This paper addresses the challenge of nasality detection in surface electromyography (EMG) based speech interfaces. We explore the existence of useful information about the velum movement and also assess if muscles deeper down in the face and neck region can be measured using surface electrodes, and the best electrode location to do so. The procedure we adopted uses Real-Time Magnetic Resonance Imaging (RT-MRI), collected from a set of speakers, providing a method to interpret EMG data. By ensuring compatible data recording conditions, and proper time alignment between the EMG and the RT-MRI data, we are able to accurately estimate the time when the velum moves and the type of movement when a nasal vowel occurs. The combination of these two sources revealed interesting and distinct characteristics in the EMG signal when a nasal vowel is uttered, which motivated a classification experiment. Overall results of this experiment provide evidence that it is possible to detect velum movement using sensors positioned below the ear, between mastoid process and the mandible, in the upper neck region. In a frame-based classification scenario, error rates as low as 32.5% for all speakers and 23.4% for the best speaker have been achieved, for nasal vowel detection. This outcome stands as an encouraging result, fostering the grounds for deeper exploration of the proposed approach as a promising route to the development of an EMG-based speech interface for languages with strong nasal characteristics.

  13. Acoustic markers of syllabic stress in Spanish excellent oesophageal speakers.

    PubMed

    Cuenca, María Heliodora; Barrio, Marina M; Anaya, Pablo; Establier, Carmelo

    2012-01-01

    The purpose of this investigation is to explore the use by Spanish excellent oesophageal speakers of acoustic cues to mark syllabic stress. The speech material has consisted of five pairs of disyllabic words which only differed in stress position. Total 44 oesophageal and 9 laryngeal speakers were recorded and a computerised designed ad hoc perceptual test was run in order to assess the accurate realisation of stress. The items produced by eight excellent oesophageal speakers with highest accuracy levels in the perception experiment were analysed acoustically with Praat, to be compared with the laryngeal control group. Measures of duration, fundamental frequency, spectral balance and overall intensity were taken for each target vowel and syllable. Results revealed that Spanish excellent oesophageal speakers were able to retain appropriate acoustic relations between stressed and unstressed syllables. Although spectral balance revealed as a strong cue for syllabic stress in the two voicing modes, a different hierarchy of acoustic cues in each voicing mode was found.

  14. Acoustic cryocooler

    DOEpatents

    Swift, Gregory W.; Martin, Richard A.; Radenbaugh, Ray

    1990-01-01

    An acoustic cryocooler with no moving parts is formed from a thermoacoustic driver (TAD) driving a pulse tube refrigerator (PTR) through a standing wave tube. Thermoacoustic elements in the TAD are spaced apart a distance effective to accommodate the increased thermal penetration length arising from the relatively low TAD operating frequency in the range of 15-60 Hz. At these low operating frequencies, a long tube is required to support the standing wave. The tube may be coiled to reduce the overall length of the cryocooler. One or two PTR's are located on the standing wave tube adjacent antinodes in the standing wave to be driven by the standing wave pressure oscillations. It is predicted that a heat input of 1000 W at 1000 K will maintian a cooling load of 5 W at 80 K.

  15. Acoustical standards in engineering acoustics

    NASA Astrophysics Data System (ADS)

    Burkhard, Mahlon D.

    2001-05-01

    The Engineering Acoustics Technical Committee is concerned with the evolution and improvement of acoustical techniques and apparatus, and with the promotion of new applications of acoustics. As cited in the Membership Directory and Handbook (2002), the interest areas include transducers and arrays; underwater acoustic systems; acoustical instrumentation and monitoring; applied sonics, promotion of useful effects, information gathering and transmission; audio engineering; acoustic holography and acoustic imaging; acoustic signal processing (equipment and techniques); and ultrasound and infrasound. Evident connections between engineering and standards are needs for calibration, consistent terminology, uniform presentation of data, reference levels, or design targets for product development. Thus for the acoustical engineer standards are both a tool for practices, for communication, and for comparison of his efforts with those of others. Development of many standards depends on knowledge of the way products are put together for the market place and acoustical engineers provide important input to the development of standards. Acoustical engineers and members of the Engineering Acoustics arm of the Society both benefit from and contribute to the Acoustical Standards of the Acoustical Society.

  16. Study of nonlinear electron-acoustic solitary and shock waves in a dissipative, nonplanar space plasma with superthermal hot electrons

    SciTech Connect

    Han, Jiu-Ning He, Yong-Lin; Luo, Jun-Hua; Nan, Ya-Gong; Han, Zhen-Hai; Dong, Guang-Xing; Duan, Wen-Shan; Li, Jun-Xiu

    2014-01-15

    With the consideration of the superthermal electron distribution, we present a theoretical investigation about the nonlinear propagation of electron-acoustic solitary and shock waves in a dissipative, nonplanar non-Maxwellian plasma comprised of cold electrons, superthermal hot electrons, and stationary ions. The reductive perturbation technique is used to obtain a modified Korteweg-de Vries Burgers equation for nonlinear waves in this plasma. We discuss the effects of various plasma parameters on the time evolution of nonplanar solitary waves, the profile of shock waves, and the nonlinear structure induced by the collision between planar solitary waves. It is found that these parameters have significant effects on the properties of nonlinear waves and collision-induced nonlinear structure.

  17. Design and analysis of the trapeziform and flat acoustic cloaks with controllable invisibility performance in a quasi-space

    NASA Astrophysics Data System (ADS)

    Zhu, Jian; Chen, Tianning; Liang, Qingxuan; Wang, Xiaopeng; Xiong, Jie; Jiang, Ping

    2015-07-01

    We present the design, implementation and detailed performance analysis for a class of trapeziform and flat acoustic cloaks. An effective large invisible area is obtained compared with the traditional carpet cloak. The cloaks are realized with homogeneous metamaterials which are made of periodic arrangements of subwavelength unit cells composed of steel embedded in air. The microstructures and its effective parameters of the cloaks are determined quickly and precisely in a broadband frequency range by using the effective medium theory and the proposed parameters optimization method. The invisibility capability of the cloaks can be controlled by the variation of the key design parameters and scale factor which are proved to have more influence on the performance in the near field than that in the far field. Different designs are suitable for different application situations. Good cloaking performance demonstrates that such a device can be physically realized with natural materials which will greatly promote the real applications of invisibility cloak.

  18. Acoustic Neuroma

    MedlinePlus

    An acoustic neuroma is a benign tumor that develops on the nerve that connects the ear to the brain. The tumor ... press against the brain, becoming life-threatening. Acoustic neuroma can be difficult to diagnose, because the symptoms ...

  19. The use of acoustic cues for phonetic identification: Effects of spectral degradation and electric hearinga)

    PubMed Central

    Winn, Matthew B.; Chatterjee, Monita; Idsardi, William J.

    2012-01-01

    Although some cochlear implant (CI) listeners can show good word recognition accuracy, it is not clear how they perceive and use the various acoustic cues that contribute to phonetic perceptions. In this study, the use of acoustic cues was assessed for normal-hearing (NH) listeners in optimal and spectrally degraded conditions, and also for CI listeners. Two experiments tested the tense/lax vowel contrast (varying in formant structure, vowel-inherent spectral change, and vowel duration) and the word-final fricative voicing contrast (varying in F1 transition, vowel duration, consonant duration, and consonant voicing). Identification results were modeled using mixed-effects logistic regression. These experiments suggested that under spectrally-degraded conditions, NH listeners decrease their use of formant cues and increase their use of durational cues. Compared to NH listeners, CI listeners showed decreased use of spectral cues like formant structure and formant change and consonant voicing, and showed greater use of durational cues (especially for the fricative contrast). The results suggest that although NH and CI listeners may show similar accuracy on basic tests of word, phoneme or feature recognition, they may be using different perceptual strategies in the process. PMID:22352517

  20. The Theory of Adaptive Dispersion and Acoustic-phonetic Properties of Cross-language Lexical-tone Systems

    NASA Astrophysics Data System (ADS)

    Alexander, Jennifer Alexandra

    Lexical-tone languages use fundamental frequency (F0/pitch) to convey word meaning. About 41.8% of the world's languages use lexical tone (Maddieson, 2008), yet those systems are under-studied. I aim to increase our understanding of speech-sound inventory organization by extending to tone-systems a model of vowel-system organization, the Theory of Adaptive Dispersion (TAD) (Liljencrants and Lindblom, 1972). This is a cross-language investigation of whether and how the size of a tonal inventory affects (A) acoustic tone-space size and (B) dispersion of tone categories within the tone-space. I compared five languages with very different tone inventories: Cantonese (3 contour, 3 level tones); Mandarin (3 contour, 1 level tone); Thai (2 contour, 3 level tones); Yoruba (3 level tones only); and Igbo (2 level tones only). Six native speakers (3 female) of each language produced 18 CV syllables in isolation, with each of his/her language's tones, six times. I measured tonal F0 across the vowel at onset, midpoint, and offglide. Tone-space size was the F0 difference in semitones (ST) between each language's highest and lowest tones. Tone dispersion was the F0 distance (ST) between two tones shared by multiple languages. Following the TAD, I predicted that languages with larger tone inventories would have larger tone-spaces. Against expectations, tone-space size was fixed across level-tone languages at midpoint and offglide, and across contour-tone languages (except Thai) at offglide. However, within each language type (level-tone vs. contour-tone), languages with smaller tone inventories had larger tone spaces at onset. Tone-dispersion results were also unexpected. The Cantonese mid-level tone was further dispersed from a tonal baseline than the Yoruba mid-level tone; Cantonese mid-level tone dispersion was therefore greater than theoretically necessary. The Cantonese high-level tone was also further dispersed from baseline than the Mandarin high-level tone -- at midpoint

  1. Acoustic Seal

    NASA Technical Reports Server (NTRS)

    Steinetz, Bruce M. (Inventor)

    2006-01-01

    The invention relates to a sealing device having an acoustic resonator. The acoustic resonator is adapted to create acoustic waveforms to generate a sealing pressure barrier blocking fluid flow from a high pressure area to a lower pressure area. The sealing device permits noncontacting sealing operation. The sealing device may include a resonant-macrosonic-synthesis (RMS) resonator.

  2. Acoustic seal

    NASA Technical Reports Server (NTRS)

    Steinetz, Bruce M. (Inventor)

    2006-01-01

    The invention relates to a sealing device having an acoustic resonator. The acoustic resonator is adapted to create acoustic waveforms to generate a sealing pressure barrier blocking fluid flow from a high pressure area to a lower pressure area. The sealing device permits noncontacting sealing operation. The sealing device may include a resonant-macrosonic-synthesis (RMS) resonator.

  3. Tongue-palate contact during selected vowels in children with speech sound disorders.

    PubMed

    Lee, Alice; Gibbon, Fiona E; Kearney, Elaine; Murphy, Doris

    2014-12-01

    There is evidence that complete tongue-palate contact across the palate during production of vowels can be observed in some children with speech disorders associated with cleft palate in the English-speaking and Japanese-speaking populations. Although it has been shown that this is not a feature of typical vowel articulation in English-speaking adults, tongue-palate contact during vowel production in typical children and English-speaking children with speech sound disorders (SSD) have not been reported in detail. Therefore, this study sought to determine whether complete tongue-palate contact occurs during production of five selected vowels in 10 children with SSD and eight typically-developing children. The results showed that none of the typical children had complete contact across the palate during any of the vowels. However, of the 119 vowels produced by the children with SSD, 24% showed complete contact across the palate during at least a portion of the vowel segment. The results from the typically-developing children suggest that complete tongue-palate contact is an atypical articulatory feature. However, the evidence suggests that this pattern occurs relatively frequently in children with SSD. Further research is needed to determine the prevalence, cause, and perceptual consequence of complete tongue-palate contact.

  4. Typological Asymmetries in Round Vowel Harmony: Support from Artificial Grammar Learning

    PubMed Central

    Finley, Sara

    2012-01-01

    Providing evidence for the universal tendencies of patterns in the world’s languages can be difficult, as it is impossible to sample all possible languages, and linguistic samples are subject to interpretation. However, experimental techniques such as artificial grammar learning paradigms make it possible to uncover the psychological reality of claimed universal tendencies. This paper addresses learning of phonological patterns (systematic tendencies in the sounds in language). Specifically, I explore the role of phonetic grounding in learning round harmony, a phonological process in which words must contain either all round vowels ([o, u]) or all unround vowels ([i, e]). The phonetic precursors to round harmony are such that mid vowels ([o, e]), which receive the greatest perceptual benefit from harmony, are most likely to trigger harmony. High vowels ([i, u]), however, are cross-linguistically less likely to trigger round harmony. Adult participants were exposed to a miniature language that contained a round harmony pattern in which the harmony source triggers were either high vowels ([i, u]) (poor harmony source triggers) or mid vowels ([o, e]) (ideal harmony source triggers). Only participants who were exposed to the ideal mid vowel harmony source triggers were successfully able to generalize the harmony pattern to novel instances, suggesting that perception and phonetic naturalness play a role in learning. PMID:23264713

  5. The discrimination and the production of English vowels by bilingual Spanish/English speakers

    NASA Astrophysics Data System (ADS)

    Levey, Sandra

    2001-05-01

    The discrimination of English vowels in real and novel words by 40 bilingual Spanish/English participants was examined. Their discrimination was compared with that of 40 native monolingual English participants. Participants were 23-36 years of age (mean 25.3; median 25.0). Stimuli were presented within triads in an ABX paradigm. This categorial discrimination paradigm was selected to avoid labeling, allowing participants to indicate categories to which stimuli belonged. Bilingual participants' productions of vowels in real words used in the discrimination task were judged by two independent listeners. The goal was to determine the degree of correlation between discrimination and production. Vowels were studied as these segments present second language learners with more difficulty than consonants. Discrimination difficulty was significantly greater for bilingual participants than for native English participants for vowel contrasts and novel words. Significant errors also appeared in the bilingual participants' production of certain vowels. English vowels absent from Spanish presented the greatest difficulty, while vowels similar to those in Spanish presented the least difficulty. Earlier age of acquisition, absence of communication problems, and greater percentage of time devoted to communication in English contributed to greater accuracy in discrimination and production. [Work supported by PSC-CUNY.

  6. Acoustic cue weighting in the singleton vs geminate contrast in Lebanese Arabic: The case of fricative consonants.

    PubMed

    Al-Tamimi, Jalal; Khattab, Ghada

    2015-07-01

    This paper is the first reported investigation of the role of non-temporal acoustic cues in the singleton-geminate contrast in Lebanese Arabic, alongside the more frequently reported temporal cues. The aim is to explore the extent to which singleton and geminate consonants show qualitative differences in a language where phonological length is prominent and where moraic structure governs segment timing and syllable weight. Twenty speakers (ten male, ten female) were recorded producing trochaic disyllables with medial singleton and geminate fricatives preceded by phonologically short and long vowels. The following acoustic measures were applied on the medial fricative and surrounding vowels: absolute duration; intensity; fundamental frequency; spectral peak and shape, dynamic amplitude, and voicing patterns of medial fricatives; and vowel quality and voice quality correlates of surrounding vowels. Discriminant analysis and receiver operating characteristics (ROC) curves were used to assess each acoustic cue's contribution to the singleton-geminate contrast. Classification rates of 89% and ROC curves with an area under the curve rate of 96% confirmed the major role played by temporal cues, with non-temporal cues contributing to the contrast but to a much lesser extent. These results confirm that the underlying contrast for gemination in Arabic is temporal, but highlight [+tense] (fortis) as a secondary feature. PMID:26233034

  7. Acoustic imaging system

    DOEpatents

    Smith, Richard W.

    1979-01-01

    An acoustic imaging system for displaying an object viewed by a moving array of transducers as the array is pivoted about a fixed point within a given plane. A plurality of transducers are fixedly positioned and equally spaced within a laterally extending array and operatively directed to transmit and receive acoustic signals along substantially parallel transmission paths. The transducers are sequentially activated along the array to transmit and receive acoustic signals according to a preestablished sequence. Means are provided for generating output voltages for each reception of an acoustic signal, corresponding to the coordinate position of the object viewed as the array is pivoted. Receptions from each of the transducers are presented on the same display at coordinates corresponding to the actual position of the object viewed to form a plane view of the object scanned.

  8. Multimode Acoustic Research

    NASA Technical Reports Server (NTRS)

    Barmatz, M.

    1985-01-01

    There is a need for high temperature containerless processing facilities that can efficiently position and manipulate molten samples in the reduced gravity environment of space. The goal of the research is to develop sophisticated high temperature manipulation capabilities such as selection of arbitrary axes rotation and rapid sample cooling. This program will investigate new classes of acoustic levitation in rectangular, cylindrical and spherical geometries. The program tasks include calculating theoretical expressions of the acoustic forces in these geometries for the excitation of up to three acoustic modes (multimodes). These calculations are used to: (1) determine those acoustic modes that produce stable levitation, (2) isolate the levitation and rotation capabilities to produce more than one axis of rotation, and (3) develop methods to translate samples down long tube cylindrical chambers. Experimental levitators will then be constructed to verify the stable levitation and rotation predictions of the models.

  9. Hyperarticulation of vowels enhances phonetic change responses in both native and non-native speakers of English: evidence from an auditory event-related potential study.

    PubMed

    Uther, Maria; Giannakopoulou, Anastasia; Iverson, Paul

    2012-08-27

    The finding that hyperarticulation of vowel sounds occurs in certain speech registers (e.g., infant- and foreigner-directed speech) suggests that hyperarticulation may have a didactic function in facilitating acquisition of new phonetic categories in language learners. This event-related potential study tested whether hyperarticulation of vowels elicits larger phonetic change responses, as indexed by the mismatch negativity (MMN) component of the auditory event-related potential (ERP) and tested native and non-native speakers of English. Data from 11 native English-speaking and 10 native Greek-speaking participants showed that Greek speakers in general had smaller MMNs compared to English speakers, confirming previous studies demonstrating sensitivity of the MMN to language background. In terms of the effect of hyperarticulation, hyperarticulated stimuli elicited larger MMNs for both language groups, suggesting vowel space expansion does elicit larger pre-attentive phonetic change responses. Interestingly Greek native speakers showed some P3a activity that was not present in the English native speakers, raising the possibility that additional attentional switch mechanisms are activated in non-native speakers compared to native speakers. These results give general support for models of speech learning such as Kuhl's Native Language Magnet enhanced (NLM-e) theory.

  10. Recognition of Devoiced Vowels Using Optical Microphone Made of Multipled POF-Type Moisture Sensors

    NASA Astrophysics Data System (ADS)

    Morisawa, Masayuki; Natori, Yoichi; Taki, Tomohito; Muto, Shinzo

    A novel optical fiber microphone system for recognizing devoiced vowels has been studied. This system consists of the optical detection of moisture pattern formed by devoiced breath and its recognization process using a modified DP-matching. To detect moisture pattern of devoiced vowels, five plastic optical fiber moisture sensors with fast response were developed and used. Using this system, high discernment rate over 93% was obtained for the devoiced vowels. This system will be used for verbally handicapped people to create sounds with a small effort in the near future.

  11. Accurate recovery of articulator positions from acoustics: New conclusions based on human data

    SciTech Connect

    Hogden, J.; Lofqvist, A.; Gracco, V.; Zlokarnik, I.; Rubin, P.; Saltzman, E.

    1996-09-01

    Vocal tract models are often used to study the problem of mapping from the acoustic transfer function to the vocal tract area function (inverse mapping). Unfortunately, results based on vocal tract models are strongly affected by the assumptions underlying the models. In this study, the mapping from acoustics (digitized speech samples) to articulation (measurements of the positions of receiver coils placed on the tongue, jaw, and lips) is examined using human data from a single speaker: Simultaneous acoustic and articulator measurements made for vowel-to-vowel transitions, /g/ closures, and transitions into and out of /g/ closures. Articulator positions were measured using an EMMA system to track coils placed on the lips, jaw, and tongue. Using these data, look-up tables were created that allow articulator positions to be estimated from acoustic signals. On a data set not used for making look-up tables, correlations between estimated and actual coil positions of around 94{percent} and root-mean-squared errors around 2 mm are common for coils on the tongue. An error source evaluation shows that estimating articulator positions from quantized acoustics gives root-mean-squared errors that are typically less than 1 mm greater than the errors that would be obtained from quantizing the articulator positions themselves. This study agrees with and extends previous studies of human data by showing that for the data studied, speech acoustics can be used to accurately recover articulator positions. {copyright} {ital 1996 Acoustical Society of America.}

  12. Space and time variations in the fine structure of the upper atmosphere according to acoustic sounding data

    NASA Astrophysics Data System (ADS)

    Kulichkov, S. N.; Chunchuzov, I. P.; Bush, G. A.; Mishenin, A. A.; Golikova, E. V.

    2016-03-01

    The results of studying variations in the fine layered structure of the upper atmosphere (heights of 20-140 km) according to data obtained from acoustic sounding within the range of infrasonic waves are given. The sources of infrasounds were surface explosions equivalent to 10 kg to 70 t of TNT. These explosions were set off in different seasons in different regions of Russia. Experimental data obtained in 1981-2011 have been analyzed. It has been found that the fine structure in the form of vertically distributed layered formations occurs in the upper atmosphere in all seasons. Moreover, the vertical distribution of both air-temperature and wind-velocity inhomogeneities in the upper atmosphere may be invariable over a time interval of no less than several hours. It has also been found that, throughout the entire atmospheric thickness from the stratopause to the lower thermosphere heights (up to 140 km), the instantaneous height distribution of layered air-temperature and wind-velocity inhomogeneities may remain almost unchanged during a time interval of no less than 20 min.

  13. Redshift-space Enhancement of Line-of-sight Baryon Acoustic Oscillations in the Sloan Digital Sky Survey Main-galaxy Sample

    NASA Astrophysics Data System (ADS)

    Tian, H. J.; Neyrinck, Mark C.; Budavári, Tamás; Szalay, Alexander S.

    2011-02-01

    We show that redshift-space distortions of galaxy correlations have a strong effect on correlation functions with distinct, localized features, like the signature of the baryon acoustic oscillations (BAO). Near the line of sight, the features become sharper as a result of redshift-space distortions. We demonstrate this effect by measuring the correlation function in Gaussian simulations and the Millennium simulation. We also analyze the SDSS DR7 main-galaxy sample, splitting the sample into slices 2fdg5 on the sky in various rotations. Measuring two-dimensional correlation functions in each slice, we do see a sharp bump along the line of sight. Using Mexican-hat wavelets, we localize it to (110 ± 10) h -1 Mpc. Averaging only along the line of sight, we estimate its significance at a particular wavelet scale and location at 2.2σ. In a flat angular weighting in the (π, rp ) coordinate system, the noise level is suppressed, pushing the bump's significance to 4σ. We estimate that there is about a 0.2% chance of getting such a signal anywhere in the vicinity of the BAO scale from a power spectrum lacking a BAO feature. However, these estimates of the significances make some use of idealized Gaussian simulations, and thus are likely a bit optimistic.

  14. Redshift-Space Enhancement of Line-of-Sight Baryon Acoustic Oscillations in the Sloan Digital Sky Survey Main-Galaxy Sample

    NASA Astrophysics Data System (ADS)

    Tian, Haijun; Neyrinck, Mark C.; Budavari, Tamas; SZALAY, AlEXANDER

    2015-08-01

    We show that redshift-space distortions of galaxy correlations have a strong effect on correlation functions with distinct, localized features, like the signature of the baryon acoustic oscillations (BAO). Near the line of sight, the features become sharper as a result of redshift-space distortions. We demonstrate this effect by measuring the correlation function in Gaussian simulations and the Millennium simulation. We also analyze the SDSS DR7 main-galaxy sample, splitting the sample into slices 2.5 on the sky in various rotations. Measuring two-dimensional correlation functions in each slice, we do see a sharp bump along the line of sight. Using Mexican-hat wavelets, we localize it to (110 ± 10) Mpc/h. Averaging only along the line of sight, we estimate its significance at a particular wavelet scale and location at 2.2σ. In a flat angular weighting in the (π,rp) coordinate system, the noise level is suppressed, pushing the bump’s significance to 4σ . We estimate that there is about a 0.2% chance of getting such a signal anywhere in the vicinity of the BAO scale from a power spectrum lacking a BAO feature. However, these estimates of the significances make some use of idealized Gaussian simulations, and thus are likely a bit optimistic.

  15. Rapid Learning of Minimally Different Words in Five- to Six-Year-Old Children: Effects of Acoustic Salience and Hearing Impairment

    ERIC Educational Resources Information Center

    Giezen, Marcel R.; Escudero, Paola; Baker, Anne E.

    2016-01-01

    This study investigates the role of acoustic salience and hearing impairment in learning phonologically minimal pairs. Picture-matching and object-matching tasks were used to investigate the learning of consonant and vowel minimal pairs in five- to six-year-old deaf children with a cochlear implant (CI), and children of the same age with normal…

  16. Quantifying Dysphonia Severity Using a Spectralcepstral-Based Acoustic Index: Comparisons with Auditory-Perceptual Judgements from the CAPE-V

    ERIC Educational Resources Information Center

    Awan, Shaheen N.; Roy, Nelson; JettE, Marie E.; Meltzner, Geoffrey S.; Hillman, Robert E.

    2010-01-01

    This study investigated the relationship between acoustic spectral/cepstral measures and listener severity ratings in normal and disordered voice samples. CAPE-V sentence samples and the vowel /[script]/were elicited from eight normal speakers and 24 patients with varying degrees of dysphonia severity. Samples were analysed for measures of the…

  17. The Effects of Noise on Speech Recognition in Cochlear Implant Subjects: Predictions and Analysis Using Acoustic Models

    NASA Astrophysics Data System (ADS)

    Remus, Jeremiah J.; Collins, Leslie M.

    2005-12-01

    Cochlear implants can provide partial restoration of hearing, even with limited spectral resolution and loss of fine temporal structure, to severely deafened individuals. Studies have indicated that background noise has significant deleterious effects on the speech recognition performance of cochlear implant patients. This study investigates the effects of noise on speech recognition using acoustic models of two cochlear implant speech processors and several predictive signal-processing-based analyses. The results of a listening test for vowel and consonant recognition in noise are presented and analyzed using the rate of phonemic feature transmission for each acoustic model. Three methods for predicting patterns of consonant and vowel confusion that are based on signal processing techniques calculating a quantitative difference between speech tokens are developed and tested using the listening test results. Results of the listening test and confusion predictions are discussed in terms of comparisons between acoustic models and confusion prediction performance.

  18. An evolution in listening: An analytical and critical study of structural, acoustic, and phenomenal aspects of selected works by Pauline Oliveros

    NASA Astrophysics Data System (ADS)

    Setar, Katherine Marie

    1997-08-01

    This dissertation analytically and critically examines composer Pauline Oliveros's philosophy of 'listening' as it applies to selected works created between 1961 and 1984. The dissertation is organized through the application of two criteria: three perspectives of listening (empirical, phenomenal, and, to a lesser extent, personal), and categories derived, in part, from her writings and interviews (improvisational, traditional, theatrical, electronic, meditational, and interactive). In general, Oliveros's works may be categorized by one of two listening perspectives. The 'empirical' listening perspective, which generally includes pure acoustic phenomenon, independent from human interpretation, is exemplified in the analyses of Sound Patterns (1961), OH HA AH (1968), and, to a lesser extent, I of IV (1966). The 'phenomenal' listening perspective, which involves the human interaction with the pure acoustic phenomenon, includes a critical examination of her post-1971 'meditation' pieces and an analytical and critical examination of her tonal 'interactive' improvisations in highly resonant space, such as Watertank Software (1984). The most pervasive element of Oliveros's stylistic evolution is her gradual change from the hierarchical aesthetic of the traditional composer, to one in which creative control is more equally shared by all participants. Other significant contributions by Oliveros include the probable invention of the 'meditation' genre, an emphasis on the subjective perceptions of musical participants as a means to greater musical awareness, her musical exploration of highly resonant space, and her pioneering work in American electronic music. Both analytical and critical commentary were applied to selective representative works from Oliveros's six compositional categories. The analytical methods applied to the Oliveros's works include Wayne Slawson's vowel/formant theory as described in his book, Sound Color, an original method of categorizing consonants as

  19. Acoustic analysis of trill sounds.

    PubMed

    Dhananjaya, N; Yegnanarayana, B; Bhaskararao, Peri

    2012-04-01

    In this paper, the acoustic-phonetic characteristics of steady apical trills--trill sounds produced by the periodic vibration of the apex of the tongue--are studied. Signal processing methods, namely, zero-frequency filtering and zero-time liftering of speech signals, are used to analyze the excitation source and the resonance characteristics of the vocal tract system, respectively. Although it is natural to expect the effect of trilling on the resonances of the vocal tract system, it is interesting to note that trilling influences the glottal source of excitation as well. The excitation characteristics derived using zero-frequency filtering of speech signals are glottal epochs, strength of impulses at the glottal epochs, and instantaneous fundamental frequency of the glottal vibration. Analysis based on zero-time liftering of speech signals is used to study the dynamic resonance characteristics of vocal tract system during the production of trill sounds. Qualitative analysis of trill sounds in different vowel contexts, and the acoustic cues that may help spotting trills in continuous speech are discussed. PMID:22501086

  20. A Psychological Experiment on the Correspondence between Colors and Voiced Vowels in Non-synesthetes'

    NASA Astrophysics Data System (ADS)

    Miyahara, Tomoko; Koda, Ai; Sekiguchi, Rikuko; Amemiya, Toshihiko

    In this study, we investigated the nature of cross-modal associations between colors and vowels. In Experiment 1, we examined the patterns of synesthetic correspondence between colors and vowels in a perceptual similarity experiment. The results were as follows: red was chosen for /a/, yellow was chosen for /i/, and blue was chosen for /o/ significantly more than any other vowels. Interestingly this pattern of correspondence is similar to the pattern of colored hearing reported by synesthetes. In Experiment 2, we investigated the robustness of these cross-modal associations using an implicit association test (IAT). A clear congruence effect was found. Participants responded faster in congruent conditions (/i/ and yellow, /o/ and blue) than in incongruent conditions (/i/ and blue, /o/ and yellow). This result suggests that the weak synesthesia between vowels and colors in non-synesthtes is not the fact of mere conscious choice, but reflects some underlying implicit associations.