首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 0 毫秒
1.
To examine the association between frequency resolution and speech recognition, auditory filter parameters and stop-consonant recognition were determined for 9 normal-hearing and 24 hearing-impaired subjects. In an earlier investigation, the relationship between stop-consonant recognition and the articulation index (AI) had been established on normal-hearing listeners. Based on AI predictions, speech-presentation levels for each subject in this experiment were selected to obtain a wide range of recognition scores. This strategy provides a method of interpreting speech-recognition performance among listeners who vary in magnitude and configuration of hearing loss by assuming that conditions which yield equal audible spectra will result in equivalent performance. It was reasoned that an association between frequency resolution and consonant recognition may be more appropriately estimated if hearing-impaired listeners' performance was measured under conditions that assured equivalent audibility of the speech stimuli. Derived auditory filter parameters indicated that filter widths and dynamic ranges were strongly associated with threshold. Stop-consonant recognition scores for most hearing-impaired listeners were not significantly poorer than predicted by the AI model. Furthermore, differences between observed recognition scores and those predicted by the AI were not associated with auditory filter characteristics, suggesting that frequency resolution and speech recognition may appear to be associated primarily because both are degraded by threshold elevation.  相似文献   

2.
Weak consonants (e.g., stops) are more susceptible to noise than vowels, owing partially to their lower intensity. This raises the question whether hearing-impaired (HI) listeners are able to perceive (and utilize effectively) the high-frequency cues present in consonants. To answer this question, HI listeners were presented with clean (noise absent) weak consonants in otherwise noise-corrupted sentences. Results indicated that HI listeners received significant benefit in intelligibility (4 dB decrease in speech reception threshold) when they had access to clean consonant information. At extremely low signal-to-noise ratio (SNR) levels, however, HI listeners received only 64% of the benefit obtained by normal-hearing listeners. This lack of equitable benefit was investigated in Experiment 2 by testing the hypothesis that the high-frequency cues present in consonants were not audible to HI listeners. This was tested by selectively amplifying the noisy consonants while leaving the noisy sonorant sounds (e.g., vowels) unaltered. Listening tests indicated small (~10%), but statistically significant, improvements in intelligibility at low SNR conditions when the consonants were amplified in the high-frequency region. Selective consonant amplification provided reliable low-frequency acoustic landmarks that in turn facilitated a better lexical segmentation of the speech stream and contributed to the small improvement in intelligibility.  相似文献   

3.
The purpose of this study was to examine the contribution of information provided by vowels versus consonants to sentence intelligibility in young normal-hearing (YNH) and typical elderly hearing-impaired (EHI) listeners. Sentences were presented in three conditions, unaltered or with either the vowels or the consonants replaced with speech shaped noise. Sentences from male and female talkers in the TIMIT database were selected. Baseline performance was established at a 70 dB SPL level using YNH listeners. Subsequently EHI and YNH participants listened at 95 dB SPL. Participants listened to each sentence twice and were asked to repeat the entire sentence after each presentation. Words were scored correct if identified exactly. Average performance for unaltered sentences was greater than 94%. Overall, EHI listeners performed more poorly than YNH listeners. However, vowel-only sentences were always significantly more intelligible than consonant-only sentences, usually by a ratio of 2:1 across groups. In contrast to written English or words spoken in isolation, these results demonstrated that for spoken sentences, vowels carry more information about sentence intelligibility than consonants for both young normal-hearing and elderly hearing-impaired listeners.  相似文献   

4.
The purpose of this study was to examine the effect of spectral-cue audibility on the recognition of stop consonants in normal-hearing and hearing-impaired adults. Subjects identified six synthetic CV speech tokens in a closed-set response task. Each syllable differed only in the initial 40-ms consonant portion of the stimulus. In order to relate performance to spectral-cue audibility, the initial 40 ms of each CV were analyzed via FFT and the resulting spectral array was passed through a sliding-filter model of the human auditory system to account for logarithmic representation of frequency and the summation of stimulus energy within critical bands. This allowed the spectral data to be displayed in comparison to a subject's sensitivity thresholds. For normal-hearing subjects, an orderly function relating the percentage of audible stimulus to recognition performance was found, with perfect discrimination performance occurring when the bulk of the stimulus spectrum was presented at suprathreshold levels. For the hearing-impaired subjects, however, it was found in many instances that suprathreshold presentation of stop-consonant spectral cues did not yield recognition equivalent to that found for the normal-hearing subjects. These results demonstrate that while the audibility of individual stop consonants is an important factor influencing recognition performance in hearing-impaired subjects, it is not always sufficient to explain the effects of sensorineural hearing loss.  相似文献   

5.
6.
The present study compared the abilities of normal and hearing-impaired subjects to discriminate differences in the spectral shapes of speechlike sounds. The minimum detectable change in amplitude of a second-formant spectral peak was determined for steady-state stimuli across a range of presentation levels. In many cases, the hearing-impaired subjects required larger spectral peaks than did the normal-hearing subjects. The performance of all subjects showed a dependence upon presentation level. For some hearing-impaired subjects, high presentation levels resulted in discrimination values similar to that of normal-hearing subjects, while for other hearing-loss subjects, increases in presentation level did not yield normal values, even when the second-formant spectral region was presented at levels above the subject's sensitivity thresholds. These results demonstrate that under certain conditions, some sensorineural hearing-impaired subjects require more prominent spectral peaks in certain speech sounds than normal subjects for equivalent performance. For the group of subjects who did not achieve normal discrimination results at any presentation level, application of high-frequency amplification to the stimuli was successful in returning those subjects' performance to within normal values.  相似文献   

7.
The purpose of this experiment was to evaluate the utilization of short-term spectral cues for recognition of initial plosive consonants (/b,d,g/) by normal-hearing and by hearing-impaired listeners differing in audiometric configuration. Recognition scores were obtained for these consonants paired with three vowels (/a,i,u/) while systematically reducing the duration (300 to 10 ms) of the synthetic consonant-vowel syllables. Results from 10 normal-hearing and 15 hearing-impaired listeners suggest that audiometric configuration interacts in a complex manner with the identification of short-duration stimuli. For consonants paired with the vowels /a/ and /u/, performance deteriorated as the slope of the audiometric configuration increased. The one exception to this result was a subject who had significantly elevated pure-tone thresholds relative to the other hearing-impaired subjects. Despite the changes in the shape of the onset spectral cues imposed by hearing loss, with increasing duration, consonant recognition in the /a/ and /u/ context for most hearing-impaired subjects eventually approached that of the normal-hearing listeners. In contrast, scores for consonants paired with /i/ were poor for a majority of hearing-impaired listeners for stimuli of all durations.  相似文献   

8.
Young normal-hearing listeners, elderly normal-hearing listeners, and elderly hearing-impaired listeners were tested on a variety of phonetic identification tasks. Where identity was cued by stimulus duration, the elderly hearing-impaired listeners evidenced normal identification functions. On a task in which there were multiple cues to vowel identity, performance was also normal. On a/b d g/identification task in which the starting frequency of the second formant was varied, performance was abnormal for both the elderly hearing-impaired listeners and the elderly normal-hearing listeners. We conclude that errors in phonetic identification among elderly hearing-impaired listeners with mild to moderate, sloping hearing impairment do not stem from abnormalities in processing stimulus duration. The results with the /b d g/continuum suggest that one factor underlying errors may be an inability to base identification on dynamic spectral information when relatively static information, which is normally characteristic of a phonetic segment, is unavailable.  相似文献   

9.
The role of different modulation frequencies in the speech envelope were studied by means of the manipulation of vowel-consonant-vowel (VCV) syllables. The envelope of the signal was extracted from the speech and the fine-structure was replaced by speech-shaped noise. The temporal envelopes in every critical band of the speech signal were notch filtered in order to assess the relative importance of different modulation frequency regions between 0 and 20 Hz. For this purpose notch filters around three center frequencies (8, 12, and 16 Hz) with three different notch widths (4-, 8-, and 12-Hz wide) were used. These stimuli were used in a consonant-recognition task in which ten normal-hearing subjects participated, and their results were analyzed in terms of recognition scores. More qualitative information was obtained with a multidimensional scaling method (INDSCAL) and sequential information analysis (SINFA). Consonant recognition is very robust for the removal of certain modulation frequency areas. Only when a wide notch around 8 Hz is applied does the speech signal become heavily degraded. As expected, the voicing information is lost, while there are different effects on plosiveness and nasality. Even the smallest filtering has a substantial effect on the transfer of the plosiveness feature, while on the other hand, filtering out only the low-modulation frequencies has a substantial effect on the transfer of nasality cues.  相似文献   

10.
Auditory filter shapes were measured for two groups of hearing-impaired subjects, young and elderly, matched for audiometric loss, for center frequencies (fc) of 100, 200, 400, and 800 Hz using a modified notched-noise method [B. R. Glasberg and B. C. J. Moore, Hear. Res. 47, 103-138 (1990)]. Two noise bands, each 0.4fc wide, were used; they were placed both symmetrically and asymmetrically about the signal frequency to allow the measurement of filter asymmetry. The overall noise level was either 77 or 87 dB SPL. Stimuli were delivered monaurally using Sennheiser HD424 earphones. Although auditory filters for the hearing-impaired subjects were generally broader than for normally hearing subjects [Moore et al., J. Acoust. Soc. Am. 87, 132-140 (1990)], some hearing-impaired subjects with mild losses had normal filters. The filters tended to broaden with increasing hearing loss. There were not any clear differences in filter characteristics between young and elderly hearing-impaired subjects. The signal-to-noise ratios at the outputs of the auditory filters required for threshold (K) tended to be lower than normal for the young hearing-impaired subjects, but were not significantly different from normal for the elderly hearing-impaired subjects. The lower K values for the young hearing-impaired subjects may occur because broadened auditory filters reduce the deleterious effects on signal detection of fluctuations in the noise.  相似文献   

11.
For a group of 30 hearing-impaired subjects and a matched group of 15 normal-hearing subjects (age range 13-17) the following data were collected: the tone audiogram, the auditory bandwidth at 1000 Hz, and the recognition threshold of a short melody presented simultaneously with two other melodies, lower and higher in frequency, respectively. The threshold was defined as the frequency distance required to recognize the test melody. It was found that, whereas the mean recognition threshold for the normal-hearing subjects was five semitones, it was, on the average, 27 semitones for the hearing-impaired subjects. Although the interindividual spread for the latter group was large, it did not correlate with the subjects' auditory bandwidth, nor with their musical experience or education.  相似文献   

12.
The influence of the precedence effect on word identification was investigated binaurally and monaurally with normally hearing and hearing-impaired subjects. The Modified Rhyme Test was processed through a PDP-12 computer to produce delay times of 0, 5, 10, 20, 40, 80, or 160 ms. The sounds were reproduced in a room by two loudspeakers positioned at +/-30 degrees azimuths in front of a subject at 50 dB SPL for normals and at the most comfortable level for impaireds. A babble of eight voices was added to reduce scores about 15% from the best values measured in quiet. Binaural and monaural word identification remained constant over a range of delays from 0 to 20 ms and declined for longer delays for both groups of subjects. The shapes of the word-identification curves were explained by self-masking (an overlap of consonants with their own repetitions) and masking (an overlap of consonants with preceding vowels or preceding and following words in sentence). Binaural responses for ten selected initial and final consonants showed various patterns of perception with delay. Some hearing impaireds showed more deterioration in word identification than others which might indicate that they experience more perceptual difficulties than normal listeners in places with reverberation or sound amplification.  相似文献   

13.
A two-alternative forced-choice task was used to measure psychometric functions for the detection of temporal gaps in a 1-kHz, 400-ms sinusoidal signal. The signal always started and finished at a positive-going zero crossing, and the gap duration was varied from 0.5 to 6.0 ms in 0.5-ms steps. The signal level was 80 dB SPL, and a spectrally shaped noise was used to mask splatter associated with the abrupt onset and offset of the signal. Two subjects with normal hearing, two subjects with unilateral cochlear hearing loss, and two subjects with bilateral cochlear hearing loss were tested. The impaired ears had confirmed reductions in frequency selectivity at 1 kHz. For the normal ears, the psychometric functions were nonmonotonic, showing minima for gap durations corresponding to integer multiples of the signal period (n ms, where n is a positive integer) and maxima for durations corresponding to (n - 0.5) ms. For the impaired ears, the psychometric functions showed only small (nonsignificant) nonmonotonicities. Performance overall was slightly worse for the impaired than for the normal ears. The main features of the results could be accounted for using a model consisting of a bandpass filter (the auditory filter), a square-law device, and a sliding temporal integrator. Consistent with the data, the model demonstrates that, although a broader auditory filter has a faster transient response, this does not necessarily lead to improved performance in a gap detection task. The model also indicates that gap thresholds do not provide a direct measure of temporal resolution, since they depend at least partly on intensity resolution.  相似文献   

14.
A significant body of evidence has accumulated indicating that vowel identification is influenced by spectral change patterns. For example, a large-scale study of vowel formant patterns showed substantial improvements in category separability when a pattern classifier was trained on multiple samples of the formant pattern rather than a single sample at steady state [J. Hillenbrand et al., J. Acoust. Soc. Am. 97, 3099-3111 (1995)]. However, in the earlier study all utterances were recorded in a constant /hVd/ environment. The main purpose of the present study was to determine whether a close relationship between vowel identity and spectral change patterns is maintained when the consonant environment is allowed to vary. Recordings were made of six men and six women producing eight vowels (see text) in isolation and in CVC syllables. The CVC utterances consisted of all combinations of seven initial consonants (/h,b,d,g,p,t,k/) and six final consonants (/b,d,g,p,t,k/). Formant frequencies for F1-F3 were measured every 5 ms during the vowel using an interactive editing tool. Results showed highly significant effects of phonetic environment. As with an earlier study of this type, particularly large shifts in formant patterns were seen for rounded vowels in alveolar environments [K. Stevens and A. House, J. Speech Hear. Res. 6, 111-128 (1963)]. Despite these context effects, substantial improvements in category separability were observed when a pattern classifier incorporated spectral change information. Modeling work showed that many aspects of listener behavior could be accounted for by a fairly simple pattern classifier incorporating F0, duration, and two discrete samples of the formant pattern.  相似文献   

15.
English consonant recognition in undegraded and degraded listening conditions was compared for listeners whose primary language was either Japanese or American English. There were ten subjects in each of the two groups, termed the non-native (Japanese) and the native (American) subjects, respectively. The Modified Rhyme Test was degraded either by a babble of voices (S/N = -3 dB) or by a room reverberation (reverberation time, T = 1.2 s). The Japanese subjects performed at a lower level than the American subjects in both noise and reverberation, although the performance difference in the undegraded, quiet condition was relatively small. There was no difference between the scores obtained in noise and in reverberation for either group. A limited-error analysis revealed some differences in type of errors for the groups of listeners. Implications of the results are discussed in terms of the effects of degraded listening conditions on non-native listeners' speech perception.  相似文献   

16.
Confusion matrices for seven synthetic steady-state vowels were obtained from ten normal and three hearing-impaired subjects. The vowels were identified at greater than 96% accuracy by the normals, and less accurately by the impaired subjects. Shortened versions of selected vowels then were used as maskers, and vowel masking patterns (VMPs) consisting of forward-masked threshold for sinusoidal probes at all vowel masker harmonics were obtained from the impaired subjects and from one normal subject. Vowel-masked probe thresholds were transformed using growth-of-masking functions obtained with flat-spectrum noise. VMPs of the impaired subjects, relative to those of the normal, were characterized by smaller dynamic range, poorer peak resolution, and poorer preservation of the vowel formant structure. These VMP characteristics, however, did not necessarily coincide with inaccurate vowel recognition. Vowel identification appeared to be related primarily to VMP peak frequencies rather than to the levels at the peaks or to between-peak characteristics of the patterns.  相似文献   

17.
18.
车型自动识别是智能交通系统的重要组成部分。针对现有车型识别存在的问题,提出利用经验模态分解和支持向量机的车型声频识别方法。将车辆行驶的声音信号进行分解,以分解不同模态的能量作为特征向量,并以此作为训练样本对支持向量机构成的车型识别器进行训练,通过对小汽车和卡车的声音信号处理结果表明:利用车辆声音信号能够正确识别不同的车型,识别准确率达95%,是车型识别的有效方法。  相似文献   

19.
This experiment measured the capability of hearing-impaired individuals to discriminate differences in the cues to the distance of spoken sentences. The stimuli were generated synthetically, using a room-image procedure to calculate the direct sound and first 74 reflections for a source placed in a 7 x 9 m room, and then presenting each of those sounds individually through a circular array of 24 loudspeakers. Seventy-seven listeners participated, aged 22-83 years and with hearing levels from -5 to 59 dB HL. In conditions where a substantial change in overall level due to the inverse-square law was available as a cue, the elderly hearing-impaired listeners did not perform any different from control groups. In other conditions where that cue was unavailable (so leaving the direct-to-reverberant relationship as a cue), either because the reverberant field dominated the direct sound or because the overall level had been artificially equalized, hearing-impaired listeners performed worse than controls. There were significant correlations with listeners' self-reported distance capabilities as measured by the "Speech, Spatial, and Qualities of Hearing" questionnaire [S. Gatehouse and W. Noble, Int. J. Audiol. 43, 85-99 (2004)]. The results demonstrate that hearing-impaired listeners show deficits in the ability to use some of the cues which signal auditory distance.  相似文献   

20.
The forward-masking properties of inharmonic complex stimuli were measured both for normal and hearing-impaired subjects. The signal threshold for a 1000-Hz pure-tone probe was obtained for six different maskers, which varied in the number of pure-tone components. The masking stimuli consisted of 1, 3, 5, 7, 9, or 11 components, logarithmically spaced in frequency surrounding the signal and presented at a fixed level of 80 dB SPL per component. In most normal-hearing subjects, the threshold for the probe decreased as the number of masking components was increased, demonstrating that stimuli with more components tended to be less effective maskers. Results from hearing-impaired subjects showed no decrease in threshold with increasing number of masking components. Instead, the thresholds increased as more components were added to the first masker. These results appear to be consistent with suppression effects within the multicomponent maskers for the normal subjects and a lack of suppression effects for the hearing-impaired subjects. The results from the normal-hearing subjects are also consistent with "across-channel" cuing.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号