首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
The perception and production of nonnative phones in second language (L2) learners can be improved via auditory training, but L2 learning is often characterized by large differences in performance across individuals. This study examined whether success in learning L2 vowels, via five sessions of high-variability phonetic training, related to the learners' native (L1) vowel processing ability or their frequency discrimination acuity. A group of native speakers of Greek received training, while another completed the pre-/post-tests but without training. Pre-/post-tests assessed different aspects of their L2 and L1 vowel processing and frequency acuity. L2 and L1 vowel processing were assessed via: (a) Natural English (L2) vowel identification in quiet and in multi-talker babble, and natural Greek (L1) vowel identification in babble; (b) the categorization of synthetic English and Greek vowel continua; and (c) discrimination of the same continua. Frequency discrimination acuity was assessed for a nonspeech continuum. Frequency discrimination acuity was related to measures of both L1 and L2 vowel processing, a finding that favors an auditory processing over a speech-specific explanation for individual variability in L2 vowel learning. The most efficient frequency discriminators at pre-test were also the most accurate both in English vowel perception and production after training.  相似文献   

2.
Recent findings in the domains of word and talker recognition reveal that listeners use previous experience with an individual talker's voice to facilitate subsequent perceptual processing of that talker's speech. These findings raise the possibility that listeners are sensitive to talker-specific acoustic-phonetic properties. The present study tested this possibility directly by examining listeners' sensitivity to talker differences in the voice-onset-time (VOT) associated with a word-initial voiceless stop consonant. Listeners were trained on the speech of two talkers. Speech synthesis was used to manipulate the VOTs of these talkers so that one had short VOTs and the other had long VOTs (counterbalanced across listeners). The results of two experiments using a paired-comparison task revealed that, when presented with a short- versus long-VOT variant of a given talker's speech, listeners could select the variant consistent with their experience of that talker's speech during training. This was true when listeners were tested on the same word heard during training and when they were tested on a different word spoken by the same talker, indicating that listeners generalized talker-specific VOT information to a novel word. Such sensitivity to talker-specific acoustic-phonetic properties may subserve at least in part listeners' capacity to benefit from talker-specific experience.  相似文献   

3.
Training Japanese listeners to identify English /r/ and /l/: a first report   总被引:5,自引:0,他引:5  
Native speakers of Japanese learning English generally have difficulty differentiating the phonemes /r/ and /l/, even after years of experience with English. Previous research that attempted to train Japanese listeners to distinguish this contrast using synthetic stimuli reported little success, especially when transfer to natural tokens containing /r/ and /l/ was tested. In the present study, a different training procedure that emphasized variability among stimulus tokens was used. Japanese subjects were trained in a minimal pair identification paradigm using multiple natural exemplars contrasting /r/ and /l/ from a variety of phonetic environments as stimuli. A pretest-posttest design containing natural tokens was used to assess the effects of training. Results from six subjects showed that the new procedure was more robust than earlier training techniques. Small but reliable differences in performance were obtained between pretest and posttest scores. The results demonstrate the importance of stimulus variability and task-related factors in training nonnative speakers to perceive novel phonetic contrasts that are not distinctive in their native language.  相似文献   

4.
Training American listeners to perceive Mandarin tones   总被引:1,自引:0,他引:1  
Auditory training has been shown to be effective in the identification of non-native segmental distinctions. In this study, it was investigated whether such training is applicable to the acquisition of non-native suprasegmental contrasts, i.c., Mandarin tones. Using the high-variability paradigm, eight American learners of Mandarin were trained in eight sessions during the course of two weeks to identify the four tones in natural words produced by native Mandarin talkers. The trainees' identification accuracy revealed an average 21% increase from the pretest to the post-test, and the improvement gained in training was generalized to new stimuli (18% increase) and to new talkers and stimuli (25% increase). Moreover, the six-month retention test showed that the improvement was retained long after training by an average 21% increase from the pretest. The results are discussed in terms of non-native suprasegmental perceptual modification, and the analogies between L2 acquisition processes at the segmental and suprasegmental levels.  相似文献   

5.
Listeners' abilities to learn to hear all the details of an initially unfamiliar sequence of ten 45-ms tones were studied by tracking detection thresholds for each tonal component over a prolonged period of training. After repeated listening to this sequence, the presence or absence of individual tones could be recognized, even though they were attenuated by 40-50 dB relative to the remainder of the pattern. Threshold-tracking histories suggest that listeners tend to employ two different learning strategies, one of which is considerably more efficient. Special training by reducing stimulus uncertainty and extending the duration of the target component was effective in increasing the rate of threshold improvement. Strategies acquired with the first pattern studied generalized to new sequences of tones. The possible implications of these results for the perceptual learning of speech or other auditory codes are discussed.  相似文献   

6.
This study investigated acoustic-phonetic correlates of intelligibility for adult and child talkers, and whether the relative intelligibility of different talkers was dependent on listener characteristics. In experiment 1, word intelligibility was measured for 45 talkers (18 women, 15 men, 6 boys, 6 girls) from a homogeneous accent group. The material consisted of 124 words familiar to 7-year-olds that adequately covered all frequent consonant confusions; stimuli were presented to 135 adult and child listeners in low-level background noise. Seven-to-eight-year-old listeners made significantly more errors than 12-year-olds or adults, but the relative intelligibility of individual talkers was highly consistent across groups. In experiment 2, listener ratings on a number of voice dimensions were obtained for the adults talkers identified in experiment 1 as having the highest and lowest intelligibility. Intelligibility was significantly correlated with subjective dimensions reflecting articulation, voice dynamics, and general quality. Finally, in experiment 3, measures of fundamental frequency, long-term average spectrum, word duration, consonant-vowel intensity ratio, and vowel space size were obtained for all talkers. Overall, word intelligibility was significantly correlated with the total energy in the 1- to 3-kHz region and word duration; these measures predicted 61% of the variability in intelligibility. The fact that the relative intelligibility of individual talkers was remarkably consistent across listener age groups suggests that the acoustic-phonetic characteristics of a talker's utterance are the primary factor in determining talker intelligibility. Although some acoustic-phonetic correlates of intelligibility were identified, variability in the profiles of the "best" talkers suggests that high intelligibility can be achieved through a combination of different acoustic-phonetic characteristics.  相似文献   

7.
8.
Some effects of talker variability on spoken word recognition   总被引:2,自引:0,他引:2  
The perceptual consequences of trial-to-trial changes in the voice of the talker on spoken word recognition were examined. The results from a series of experiments using perceptual identification and naming tasks demonstrated that perceptual performance decreases when the voice of the talker changes from trial to trial compared to performance when the voice on each trial remains the same. In addition, the effects of talker variability on word recognition appeared to be more robust and less dependent on task than the effects of word frequency and lexical structure. Possible hypotheses regarding the nature of the processes giving rise to these effects are discussed, with particular attention to the idea that the processing of information about the talker's voice is intimately related to early perceptual processes that extract acoustic-phonetic information from the speech signal.  相似文献   

9.
Adaptation to the acoustic world following cochlear implantation does not typically include formal training or extensive audiological rehabilitation. Can cochlear implant (CI) users benefit from formal training, and if so, what type of training is best? This study used a pre-/posttest design to evaluate the efficacy of training and generalization of perceptual learning in normal hearing subjects listening to CI simulations (eight-channel sinewave vocoder). Five groups of subjects were trained on words (simple/complex), sentences (meaningful/anomalous), or environmental sounds, and then were tested using an open-set identification task. Subjects were trained on only one set of materials but were tested on all stimuli. All groups showed significant improvement due to training, which successfully generalized to some, but not all stimulus materials. For easier tasks, all types of training generalized equally well. For more difficult tasks, training specificity was observed. Training on speech did not generalize to the recognition of environmental sounds; however, explicit training on environmental sounds successfully generalized to speech. These data demonstrate that the perceptual learning of degraded speech is highly context dependent and the type of training and the specific stimulus materials that a subject experiences during perceptual learning has a substantial impact on generalization to new materials.  相似文献   

10.
While numerous studies on infant perception have demonstrated the infant's ability to discriminate sounds having different frequencies, little research has evaluated more sophisticated pitch perception abilities such as perceptual constancy and perception of the missing fundamental. In the present study 7-8-month-old infants demonstrated the ability to discriminate harmonic complexes from two pitch categories that differed in pitch by approximately 20% (e.g., 160 vs 200 Hz). Using a visually reinforced conditioned head-turning paradigm, a number of spectrally different tonal complexes that contained varying harmonic components but signaled the same two pitch categories were presented. After learning the basic pitch discrimination, the same infants learned to categorize spectrally different tonal complexes according to the pitches signaled by their fundamental frequencies. That is, the infants showed evidence of perceptual constancy for the pitch of harmonic complexes. Finally, infants heard tonal complexes that signaled the same pitch categories but for which the fundamental frequency was removed. Infants were still able to categorize the harmonic complexes according to their pitch categories. These results suggest that by 7 months of age infants show fairly sophisticated pitch perception abilities similar to those demonstrated by adults.  相似文献   

11.
The speech signal may be divided into frequency bands, each containing temporal properties of the envelope and fine structure. For maximal speech understanding, listeners must allocate their perceptual resources to the most informative acoustic properties. Understanding this perceptual weighting is essential for the design of assistive listening devices that need to preserve these important speech cues. This study measured the perceptual weighting of young normal-hearing listeners for the envelope and fine structure in each of three frequency bands for sentence materials. Perceptual weights were obtained under two listening contexts: (1) when each acoustic property was presented individually and (2) when multiple acoustic properties were available concurrently. The processing method was designed to vary the availability of each acoustic property independently by adding noise at different levels. Perceptual weights were determined by correlating a listener's performance with the availability of each acoustic property on a trial-by-trial basis. Results demonstrated that weights were (1) equal when acoustic properties were presented individually and (2) biased toward envelope and mid-frequency information when multiple properties were available. Results suggest a complex interaction between the available acoustic properties and the listening context in determining how best to allocate perceptual resources when listening to speech in noise.  相似文献   

12.
A group of prelinguistically hearing impaired children, between 7 and 11 years of age, were trained in the perception of vowel duration and place, the fricative /s/, and manner of articulation (/m/ vs /b/ and /s/ vs /t/) distinctions, using information provided by a multiple-channel electrotactile aid (Tickle Talker), and through aided hearing. Training was provided in the tactile-plus-aided hearing (TA) and tactile (T) conditions. Speech feature recognition tests were conducted in the TA, T, and aided hearing (A) conditions, during pretraining, training, and post-training phases. Test scores in the TA and T conditions were significantly greater than scores in the A condition for all tests, suggesting that perception of these features was improved when the tactile aid was worn. Test scores in the training and post-training phases were significantly greater than in the pretraining phase, suggesting that the training provided was responsible for the improvement in feature perception. Statistical analyses demonstrated a significant interaction between the main effects of condition and phase, suggesting that training improved perception in the TA and T conditions, but not in the A condition. Post-training and training test scores were similar suggesting that the perceptual skills acquired during training were retained after the removal of training. Recognition of trained features improved for trained, as well as for untrained words.  相似文献   

13.
The purpose of this exploratory study was to examine the relationship between undergraduate vocal music majors' diction acquisition abilities for singing in a nonnative language (as rated both by themselves and by their studio voice teachers) and their scores on an objective test of phonemic and stress perception. Ten students with varying levels of university voice training served as participants. The results showed significant negative correlations between each of the teachers' four ratings and the students' scores on the phonemic awareness subtest. In addition, 20% of the students demonstrated evidence of underdeveloped phonemic awareness skills, as indicated by their below average test performance. Considerable individual differences were also observed in the students' abilities to track phonemes within a sequence of phonemes, count and track syllables within a sequence of syllables, and track combinations of phoneme and syllable changes in sequence, as evidenced by subtest performance scores. These findings corroborate existing reports which indicate that approximately 30% of the population does not fully develop phonemic awareness skills in the absence of special training. The findings support the utility of this objective test of phonemic and stress perception as a means of identifying students who will have difficulty with diction acquisition, and point to possibilities for pretraining to improve their response to diction instruction.  相似文献   

14.
Three experiments were conducted to examine the effects of trial-to-trial variations in speaking style, fundamental frequency, and speaking rate on identification of spoken words. In addition, the experiments investigated whether any effects of stimulus variability would be modulated by phonetic confusability (i.e., lexical difficulty). In Experiment 1, trial-to-trial variations in speaking style reduced the overall identification performance compared with conditions containing no speaking-style variability. In addition, the effects of variability were greater for phonetically confusable words than for phonetically distinct words. In Experiment 2, variations in fundamental frequency were found to have no significant effects on spoken word identification and did not interact with lexical difficulty. In Experiment 3, two different methods for varying speaking rate were found to have equivalent negative effects on spoken word recognition and similar interactions with lexical difficulty. Overall, the findings are consistent with a phonetic-relevance hypothesis, in which accommodating sources of acoustic-phonetic variability that affect phonetically relevant properties of speech signals can impair spoken word identification. In contrast, variability in parameters of the speech signal that do not affect phonetically relevant properties are not expected to affect overall identification performance. Implications of these findings for the nature and development of lexical representations are discussed.  相似文献   

15.
In English, voiced and voiceless syllable-initial stop consonants differ in both fundamental frequency at the onset of voicing (onset F0) and voice onset time (VOT). Although both correlates, alone, can cue the voicing contrast, listeners weight VOT more heavily when both are available. Such differential weighting may arise from differences in the perceptual distance between voicing categories along the VOT versus onset F0 dimensions, or it may arise from a bias to pay more attention to VOT than to onset F0. The present experiment examines listeners' use of these two cues when classifying stimuli in which perceptual distance was artificially equated along the two dimensions. Listeners were also trained to categorize stimuli based on one cue at the expense of another. Equating perceptual distance eliminated the expected bias toward VOT before training, but successfully learning to base decisions more on VOT and less on onset F0 was easier than vice versa. Perceptual distance along both dimensions increased for both groups after training, but only VOT-trained listeners showed a decrease in Garner interference. Results lend qualified support to an attentional model of phonetic learning in which learning involves strategic redeployment of selective attention across integral acoustic cues.  相似文献   

16.
Pure tone intensity discrimination thresholds can be elevated by the introduction of remote maskers with roved level. This effect is on the order of 10 dB [10 log(DeltaII)] in some conditions and can be demonstrated under conditions of little or no energetic masking. The current study examined the effect of practice and observer strategy on this phenomenon. Experiment 1 included observers who had no formal experience with intensity discrimination and provided training over 6 h on a single masked intensity discrimination task to assess learning effects. Thresholds fell with practice for most observers, with significant improvements in six out of eight cases. Despite these improvements significant masking remained in all cases. The second experiment assessed trial-by-trial effects of roved masker level. Conditional probability of a "signal-present" response as a function of the rove value assigned to each of the two masker tones indicates fundamental differences among observers' processing strategies, even after 6 h of practice. The variability in error patterns across practiced listeners suggests that observers approach the task differently, though this variability does not appear to be related to sensitivity.  相似文献   

17.
Within the debate on the mechanisms underlying infants' perceptual acquisition, one hypothesis proposes that infants' perception is directly affected by the acoustic implementation of sound categories in the speech they hear. In consonance with this view, the present study shows that individual variation in fine-grained, subphonemic aspects of the acoustic realization of /s/ in caregivers' speech predicts infants' discrimination of this sound from the highly similar /∫/, suggesting that learning based on acoustic cue distributions may indeed drive natural phonological acquisition.  相似文献   

18.
Learning to recognize complex sensory signals can change the way they are perceived. European starlings (Sturnus vulgaris) recognize other starlings by their song, which consists of a series of complex, stereotyped motifs. Song recognition learning is accompanied by plasticity in secondary auditory areas, suggesting that perceptual learning is involved. Here, to investigate whether perceptual learning can be observed behaviorally, a same-different operant task was used to measure how starlings perceived small differences in motif structure. Birds trained to recognize conspecific songs were better at detecting variations in motifs from the songs they learned, even though this variation was not directly necessary to learn the associative task. Discrimination also improved as the reference stimulus was repeated multiple times. Perception of the much larger differences between different motifs was unaffected by training. These results indicate that sensory representations of motifs are enhanced when starlings learn to recognize songs.  相似文献   

19.
This study examined perceptual learning of spectrally complex nonspeech auditory categories in an interactive multi-modal training paradigm. Participants played a computer game in which they navigated through a three-dimensional space while responding to animated characters encountered along the way. Characters' appearances in the game correlated with distinctive sound category distributions, exemplars of which repeated each time the characters were encountered. As the game progressed, the speed and difficulty of required tasks increased and characters became harder to identify visually, so quick identification of approaching characters by sound patterns was, although never required or encouraged, of gradually increasing benefit. After 30 min of play, participants performed a categorization task, matching sounds to characters. Despite not being informed of audio-visual correlations, participants exhibited reliable learning of these patterns at posttest. Categorization accuracy was related to several measures of game performance and category learning was sensitive to category distribution differences modeling acoustic structures of speech categories. Category knowledge resulting from the game was qualitatively different from that gained from an explicit unsupervised categorization task involving the same stimuli. Results are discussed with respect to information sources and mechanisms involved in acquiring complex, context-dependent auditory categories, including phonetic categories, and to multi-modal statistical learning.  相似文献   

20.
Speech perception requires the integration of information from multiple phonetic and phonological dimensions. A sizable literature exists on the relationships between multiple phonetic dimensions and single phonological dimensions (e.g., spectral and temporal cues to stop consonant voicing). A much smaller body of work addresses relationships between phonological dimensions, and much of this has focused on sequences of phones. However, strong assumptions about the relevant set of acoustic cues and/or the (in)dependence between dimensions limit previous findings in important ways. Recent methodological developments in the general recognition theory framework enable tests of a number of these assumptions and provide a more complete model of distinct perceptual and decisional processes in speech sound identification. A hierarchical Bayesian Gaussian general recognition theory model was fit to data from two experiments investigating identification of English labial stop and fricative consonants in onset (syllable initial) and coda (syllable final) position. The results underscore the importance of distinguishing between conceptually distinct processing levels and indicate that, for individual subjects and at the group level, integration of phonological information is partially independent with respect to perception and that patterns of independence and interaction vary with syllable position.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号