首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 31 毫秒
1.
A set of experiments was conducted to examine the loudness of sounds with temporally asymmetric amplitude envelopes. Envelopes were generated with fast-attack/slow-decay characteristics to produce F-S (or "fast-slow") stimuli, while temporally reversed versions of these same envelopes produced corresponding S-F ("slow-fast") stimuli. For sinusoidal (330-6000 Hz) and broadband noise carriers, S-F stimuli were louder than F-S stimuli of equal energy. The magnitude of this effect was sensitive to stimulus order, with the largest differences between F-S and S-F loudness occurring after exposure to a preceding F-S stimulus. These results are not compatible with automatic gain control, power-spectrum models of loudness, or predictions obtained using the auditory image model [Patterson et al., J. Acoust. Soc. Am. 98, 1890-1894 (1995)]. Rather, they are comparable to phenomena of perceptual constancy, and may be related to the parsing of auditory input into direct and reverberant sound.  相似文献   

2.
Mammals have evolved the ability to acquire auditory discriminations. The characteristics of this discriminative ability presumably fit the natural conditions under which discriminations are normally acquired. The purpose of this paper is to review experiments which were directed at showing that auditory discriminations are most rapidly acquired when natural features are incorporated into the experiments. The experiments were also directed at discovering the underlying characteristics of the discriminative ability. When animals were trained to discriminate the position of a sound source in which natural features were incorporated into the experiment, the discrimination was acquired in one trial. Manipulation of the natural features suggested that one trial acquisition depends upon the following. (1) Stimulus novelty; the effect of reinforcement is stronger in the presence of novel than familiar stimuli. (2) Specific behavioral effect of reinforcement; the effect of reinforcing a response in the presence of a novel auditory stimulus is to increase the strength of approaching and manipulating the sound source.  相似文献   

3.
Temporal resolution is often measured using the detection of temporal gaps or signals in temporal gaps embedded in long-duration stimuli. In this study, psychoacoustical paradigms are developed for measuring the temporal encoding of transient stimuli. The stimuli consisted of very short pips which, in two experiments, contained a steady state portion. The carrier was high-pass filtered, dynamically compressed noise, refreshed for every stimulus presentation. The first experiment shows that, with these very short stimuli, gap detection thresholds are about the same as obtained in previous investigations. Experiments II and III show that, using the same stimuli, temporal-separation thresholds and duration-discrimination thresholds are better than gap-detection thresholds. Experiment IV investigates the significance of residual spectral cues for the listeners' performance. In experiment V, temporal separation thresholds were measured as a function of the signal-pip sensation level (SL) in both forward- and backward-masking conditions. The separation thresholds show a strong temporal asymmetry with good separation thresholds independent of signal-pip SL in backward-masking conditions and increasing separation thresholds with decreasing signal-pip SL in forward-masking conditions. A model of the auditory periphery is used to stimulate the gap-detection and temporal-separation thresholds quantitatively. By varying parameters like auditory-filter width and transduction time constants, the model provides some insight into how the peripheral auditory system may cope with temporal processing tasks and thus represents a more physiology-related complement to current models of temporal processing.  相似文献   

4.
Phonemic and phonetic factors in adult cross-language speech perception   总被引:5,自引:0,他引:5  
Previous research has indicated that young infants can discriminate speech sounds across phonetic boundaries regardless of specific relevant experience, and that there is a modification in this ability during ontogeny such that adults often have difficulty discriminating phonetic contrasts which are not used contrastively in their native language. This pattern of findings has often been interpreted as suggesting that humans are endowed with innate auditory sensitivities which enable them to discriminate speech sounds according to universal phonetic boundaries and that there is a decline or loss in this ability after being exposed to a language which contrasts only a subset of those distinctions. The present experiments were designed to determine whether this modification represents a loss of sensorineural response capabilities or whether it shows a shift in attentional focus and/or processing strategies. In experiment 1, adult English-speaking subjects were tested on their ability to discriminate two non-English speech contrasts in a category-change discrimination task after first being predisposed to adopt one of four perceptual sets. In experiments 2, 3, and 4 subjects were tested in an AX (same/different) procedure, and the effects of both limited training and duration of the interstimulus interval were assessed. Results suggest that the previously observed ontogenetic modification in the perception of non-native phonetic contrasts involves a change in processing strategies rather than a sensorineural loss. Adult listeners can discriminate sounds across non-native phonetic categories in some testing conditions, but are not able to use that ability in testing conditions which have demands similar to those required in natural language processing.  相似文献   

5.
Loudness functions and frequency difference limens (DLFs) were measured in five subjects with steeply sloping high-frequency sensorineural hearing loss. The stimuli were pulsed pure tones encompassing a range of frequencies. Loudness data were obtained using a 2AFC matching procedure with a 500-Hz reference presented at a number of levels. DLFs were measured using a 3AFC procedure with intensities randomized within 6 dB around an equal-loudness level. Results showed significantly shallower loudness functions near the cutoff frequency of the loss than at a lower frequency, where hearing thresholds were near normal. DLFs were elevated, on average, relative to DLFs measured using the same procedure in five normally hearing subjects, but showed a local reduction near the cutoff frequency in most subjects with high-frequency loss. The loudness data are generally consistent with recent models that describe loudness perception in terms of peripheral excitation patterns that are presumably restricted by a steeply sloping hearing loss. However, the DLF data are interpreted with reference to animal experiments that have shown reorganization in the auditory cortex following the introduction of restricted cochlear lesions. Such reorganization results in an increase in the spatial representation of lesion-edge frequencies, and is comparable with the functional reorganization observed in animals following frequency-discrimination training. It is suggested that similar effects may occur in humans with steeply sloping high-frequency hearing loss, and therefore, the local reduction in DLFs in our data may reflect neural plasticity.  相似文献   

6.
This study examined the perceptual specialization for native-language speech sounds, by comparing native Hindi and English speakers in their perception of a graded set of English /w/-/v/ stimuli that varied in similarity to natural speech. The results demonstrated that language experience does not affect general auditory processes for these types of sounds; there were strong cross-language differences for speech stimuli, and none for stimuli that were nonspeech. However, the cross-language differences extended into a gray area of speech-like stimuli that were difficult to classify, suggesting that the specialization occurred in phonetic processing prior to categorization.  相似文献   

7.
8.
9.
The cortical mechanisms of perceptual segregation of concurrent sound sources were examined, based on binaural detection of interaural timing differences. Auditory event-related potentials were measured from 11 healthy subjects. Binaural stimuli were created by introducing a dichotic delay of 500-ms duration to a narrow frequency region within a broadband noise, and resulted in a perception of a centrally located noise and a right-lateralized pitch (dichotic pitch). In separate listening conditions, subjects actively discriminated and responded to randomly interleaved binaural and control stimuli, or ignored random stimuli while watching silent cartoons. In a third listening condition subjects ignored stimuli presented in homogenous blocks. For all listening conditions, the dichotic pitch stimulus elicited an object-related negativity (ORN) at a latency of about 150-250 ms after stimulus onset. When subjects were required to actively respond to stimuli, the ORN was followed by a P400 wave with a latency of about 320-420 ms. These results support and extend a two-stage model of auditory scene analysis in which acoustic streams are automatically parsed into component sound sources based on source-relevant cues, followed by a controlled process involving identification and generation of a behavioral response.  相似文献   

10.
The perceptual mechanisms of assimilation and contrast in the phonetic perception of vowels were investigated. In experiment 1, 14 stimulus continua were generated using an /i/-/e/-/a/ vowel continuum. They ranged from a continuum with both ends belonging to the same phonemic category in Japanese, to a continuum with both ends belonging to different phonemic categories. The AXB method was employed and the temporal position of X was changed under three conditions. In each condition ten subjects were required to judge whether X was similar to A or to B. The results demonstrated that assimilation to the temporally closer sound occurs if the phonemic categories of A and B are the same and that contrast to the temporally closer sound occurs if A and B belong to different phonemic categories. It was observed that the transition from assimilation to contrast is continuous except in the /i'/-X-/e/ condition. In experiment 2, the total duration of t 1 (between A and X) and t 2 (between X and B) was changed under five conditions. One stimulus continuum consisted of the same phonemic category in Japanese and the other consisted of different phonemic categories. Six subjects were required to make similarity judgements of X. The results demonstrated that the occurrence of assimilation and contrast to the temporally closer sound seemed to be constant under each of the five conditions. The present findings suggest that assimilation and contrast are determined by three factors: the temporal position of the three stimuli, the acoustic distance between the three stimuli on the stimulus continuum, and the phonemic categories of the three stimuli.  相似文献   

11.
Perceptual integration of vibrotactile and auditory sinusoidal tone pulses was studied in detection experiments as a function of stimulation frequency. Vibrotactile stimuli were delivered through a single channel vibrator to the left middle fingertip. Auditory stimuli were presented diotically through headphones in a background of 50 dB sound pressure level broadband noise. Detection performance for combined auditory-tactile presentations was measured using stimulus levels that yielded 63% to 77% correct unimodal performance. In Experiment 1, the vibrotactile stimulus was 250 Hz and the auditory stimulus varied between 125 and 2000 Hz. In Experiment 2, the auditory stimulus was 250 Hz and the tactile stimulus varied between 50 and 400 Hz. In Experiment 3, the auditory and tactile stimuli were always equal in frequency and ranged from 50 to 400 Hz. The highest rates of detection for the combined-modality stimulus were obtained when stimulating frequencies in the two modalities were equal or closely spaced (and within the Pacinian range). Combined-modality detection for closely spaced frequencies was generally consistent with an algebraic sum model of perceptual integration; wider-frequency spacings were generally better fit by a Pythagorean sum model. Thus, perceptual integration of auditory and tactile stimuli at near-threshold levels appears to depend both on absolute frequency and relative frequency of stimulation within each modality.  相似文献   

12.
The role of auditory feedback in speech motor control was explored in three related experiments. Experiment 1 investigated auditory sensorimotor adaptation: the process by which speakers alter their speech production to compensate for perturbations of auditory feedback. When the first formant frequency (F1) was shifted in the feedback heard by subjects as they produced vowels in consonant-vowel-consonant (CVC) words, the subjects' vowels demonstrated compensatory formant shifts that were maintained when auditory feedback was subsequently masked by noise-evidence of adaptation. Experiment 2 investigated auditory discrimination of synthetic vowel stimuli differing in F1 frequency, using the same subjects. Those with more acute F1 discrimination had compensated more to F1 perturbation. Experiment 3 consisted of simulations with the directions into velocities of articulators model of speech motor planning, which showed that the model can account for key aspects of compensation. In the model, movement goals for vowels are regions in auditory space; perturbation of auditory feedback invokes auditory feedback control mechanisms that correct for the perturbation, which in turn causes updating of feedforward commands to incorporate these corrections. The relation between speaker acuity and amount of compensation to auditory perturbation is mediated by the size of speakers' auditory goal regions, with more acute speakers having smaller goal regions.  相似文献   

13.
Although in a number of experiments noise-band vocoders have been shown to provide acoustic models for speech perception in cochlear implants (CI), the present study assesses in four experiments whether and under what limitations noise-band vocoders can be used as an acoustic model for pitch perception in CI. The first two experiments examine the effect of spectral smearing on simulated electrode discrimination and fundamental frequency (FO) discrimination. The third experiment assesses the effect of spectral mismatch in an FO-discrimination task with two different vocoders. The fourth experiment investigates the effect of amplitude compression on modulation rate discrimination. For each experiment, the results obtained from normal-hearing subjects presented with vocoded stimuli are compared to results obtained directly from CI recipients. The results show that place pitch sensitivity drops with increased spectral smearing and that place pitch cues for multi-channel stimuli can adequately be mimicked when the discriminability of adjacent channels is adjusted by varying the spectral slopes to match that of CI subjects. The results also indicate that temporal pitch sensitivity is limited for noise-band carriers with low center frequencies and that the absence of a compression function in the vocoder might alter the saliency of the temporal pitch cues.  相似文献   

14.
The accuracy of a method of simulating reduced frequency selectivity by the spectral smearing of complex stimuli has been evaluated. First an excitation pattern that would be evoked by a given nonsmeared stimulus in an impaired ear with broad auditory filters was estimated. Then the spectral smearing of the stimulus that would be necessary to create the same excitation pattern in a normal ear was calculated. The smearing was based on the shapes of simulated broad auditory filters; both symmetric and asymmetric broad filters were simulated. The method was used to process notched noise, and tones in notched noise, and the processed stimuli were used in a series of experiments with normally hearing subjects measuring the threshold for the tone in notched noise. The resulting data were used to derive auditory filter shapes. The derived filter shapes were generally similar to the expected shapes (based on the type of spectral smearing used), but there were some systematic discrepancies and some individual differences. The discrepancies do not seem to be due to the use of information derived from phase locking, since they were observed both at 1 kHz (where phase locking occurs) and at 6 kHz (where phase locking probably does not occurs). The discrepancies also do not seem to be due to the transmission characteristics of the outer/middle ear, since they occurred both when these characteristics were taken into account in the fitting procedure, and when the stimuli were preshaped to compensate for these characteristics. The influence of the subjects' own auditory filters probably can explain some of the discrepancies; the excitation pattern evoked by the spectrally smeared stimuli can be significantly influenced by the subjects' own filters when those filters are not much sharper than the simulated filters used to produce the smeared stimuli. Finally, some of the discrepancies can probably be explained by subjects combining information across auditory filters, rather than just using the single 'best' filter in each condition; this represents a limitation of the fitting procedure rather than of the simulation itself. Overall, the simulation worked reasonably well, especially when the smearing was based on symmetric filters.  相似文献   

15.
Spectral integration refers to the summation of activity beyond the bandwidth of the peripheral auditory filter. Several experimental lines have sought to determine the bandwidth of this "supracritical" band phenomenon. This paper reports on two experiments which tested the limit on spectral integration in the same listeners. Experiment I verified the critical separation of 3.5 bark in two-formant synthetic vowels as advocated by the center-of-gravity (COG) hypothesis. According to the COG effect, two formants are integrated into a single perceived peak if their separation does not exceed approximately 3.5 bark. With several modifications to the methods of a classic COG matching task, the present listeners responded to changes in pitch in two-formant synthetic vowels, not estimating their phonetic quality. By changing the amplitude ratio of the formants, the frequency of the perceived peak was closer to that of the stronger formant. This COG effect disappeared with larger formant separation. In a second experiment, auditory spectral resolution bandwidths were measured for the same listeners using common-envelope, two-tone complex signals. Results showed that the limits of spectral averaging in two-formant vowels and two-tone spectral resolution bandwidth were related for two of the three listeners. The third failed to perform the discrimination task. For the two subjects who completed both tasks, the results suggest that the critical region in vowel task and the complex-tone discriminability estimates are linked to a common mechanism, i.e., to an auditory spectral resolving power. A signal-processing model is proposed to predict the COG effect in two-formant synthetic vowels. The model introduces two modifications to Hermansky's [J. Acoust. Soc. Am. 87, 1738-1752 (1990)] perceptual linear predictive (PLP) model. The model predictions are generally compatible with the present experimental results and with the predictions of several earlier models accounting for the COG effect.  相似文献   

16.
Recent work [Iverson et al. (2003) Cognition, 87, B47-57] has suggested that Japanese adults have difficulty learning English /r/ and /l/ because they are overly sensitive to acoustic cues that are not reliable for /r/-/l/ categorization (e.g., F2 frequency). This study investigated whether cue weightings are altered by auditory training, and compared the effectiveness of different training techniques. Separate groups of subjects received High Variability Phonetic Training (natural words from multiple talkers), and 3 techniques in which the natural recordings were altered via signal processing (All Enhancement, with F3 contrast maximized and closure duration lengthened; Perceptual Fading, with F3 enhancement reduced during training; and Secondary Cue Variability, with variation in F2 and durations increased during training). The results demonstrated that all of the training techniques improved /r/-/l/ identification by Japanese listeners, but there were no differences between the techniques. Training also altered the use of secondary acoustic cues; listeners became biased to identify stimuli as English /l/ when the cues made them similar to the Japanese /r/ category, and reduced their use of secondary acoustic cues for stimuli that were dissimilar to Japanese /r/. The results suggest that both category assimilation and perceptual interference affect English /r/ and /l/ acquisition.  相似文献   

17.
Both dyslexics and auditory neuropathy (AN) subjects show inferior consonant-vowel (CV) perception in noise, relative to controls. To better understand these impairments, natural acoustic speech stimuli that were masked in speech-shaped noise at various intensities were presented to dyslexic, AN, and control subjects either in isolation or accompanied by visual articulatory cues. AN subjects were expected to benefit from the pairing of visual articulatory cues and auditory CV stimuli, provided that their speech perception impairment reflects a relatively peripheral auditory disorder. Assuming that dyslexia reflects a general impairment of speech processing rather than a disorder of audition, dyslexics were not expected to similarly benefit from an introduction of visual articulatory cues. The results revealed an increased effect of noise masking on the perception of isolated acoustic stimuli by both dyslexic and AN subjects. More importantly, dyslexics showed less effective use of visual articulatory cues in identifying masked speech stimuli and lower visual baseline performance relative to AN subjects and controls. Last, a significant positive correlation was found between reading ability and the ameliorating effect of visual articulatory cues on speech perception in noise. These results suggest that some reading impairments may stem from a central deficit of speech processing.  相似文献   

18.
Listeners' abilities to learn to hear all the details of an initially unfamiliar sequence of ten 45-ms tones were studied by tracking detection thresholds for each tonal component over a prolonged period of training. After repeated listening to this sequence, the presence or absence of individual tones could be recognized, even though they were attenuated by 40-50 dB relative to the remainder of the pattern. Threshold-tracking histories suggest that listeners tend to employ two different learning strategies, one of which is considerably more efficient. Special training by reducing stimulus uncertainty and extending the duration of the target component was effective in increasing the rate of threshold improvement. Strategies acquired with the first pattern studied generalized to new sequences of tones. The possible implications of these results for the perceptual learning of speech or other auditory codes are discussed.  相似文献   

19.
Adaptation to the acoustic world following cochlear implantation does not typically include formal training or extensive audiological rehabilitation. Can cochlear implant (CI) users benefit from formal training, and if so, what type of training is best? This study used a pre-/posttest design to evaluate the efficacy of training and generalization of perceptual learning in normal hearing subjects listening to CI simulations (eight-channel sinewave vocoder). Five groups of subjects were trained on words (simple/complex), sentences (meaningful/anomalous), or environmental sounds, and then were tested using an open-set identification task. Subjects were trained on only one set of materials but were tested on all stimuli. All groups showed significant improvement due to training, which successfully generalized to some, but not all stimulus materials. For easier tasks, all types of training generalized equally well. For more difficult tasks, training specificity was observed. Training on speech did not generalize to the recognition of environmental sounds; however, explicit training on environmental sounds successfully generalized to speech. These data demonstrate that the perceptual learning of degraded speech is highly context dependent and the type of training and the specific stimulus materials that a subject experiences during perceptual learning has a substantial impact on generalization to new materials.  相似文献   

20.
The influence of different memory systems and associated attentional processes on the acuity of auditory images, formed for the purpose of making intonation judgments, was examined across three experiments using three different task types (cued-attention, imagery, and two-tone discrimination). In experiment 1 the influence of implicit long-term memory for musical scale structure was manipulated by varying the scale degree (leading tone versus tonic) of the probe note about which a judgment had to be made. In experiments 2 and 3 the ability of short-term absolute pitch knowledge to develop was manipulated by presenting blocks of trials in the same key or in seven different keys. The acuity of auditory images depended on all of these manipulations. Within individual listeners, thresholds in the two-tone discrimination and cued-attention conditions were closely related. In many listeners, cued-attention thresholds were similar to thresholds in the imagery condition, and depended on the amount of training individual listeners had in playing a musical instrument. The results indicate that mental images formed at a sensory/cognitive interface for the purpose of making perceptual decisions are highly malleable.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号