首页 | 本学科首页   官方微博 | 高级检索  
相似文献
 共查询到20条相似文献,搜索用时 15 毫秒
1.
An experiment was designed to determine whether normally sighted human subjects would be able to adapt to the handicapping effects of sudden deprivation of visual cues on horizontal plane sound localization. Two groups of sighted normal-hearing young adults participated. One group was allowed the benefit of sight. The other group was blindfolded. Measurements of accuracy and the time to respond were made daily over the course of five consecutive days, in a semi-reverberant sound proof booth that modeled listening in a small office. Sound localization was assessed using an array of eight speakers that surrounded the subject in space. Each day, one block of 120 trials was presented for each of three stimuli, two one-third octave noise bands, centred at 0.5 and 4 kHz, and broadband noise, to assess the utilization of interaural temporal difference cues, interaural level difference cues and binaural and spectral cues in combination. Blindfolded subjects were relatively less accurate than sighted subjects. Both groups showed gains with practice, the blindfolded group to a greater degree, largely due to improvements in the use of spectral cues. The blindfolded group took longer to respond than the sighted group, but showed greater decrements in response time with practice.  相似文献   

2.
An experiment was conducted to determine the effect of aging on sound localization. Seven groups of 16 subjects, aged 10-81 years, were tested. Sound localization was assessed using six different arrays of four or eight loudspeakers that surrounded the subject in the horizontal plane, at a distance of 1 m. For two 4-speaker arrays, one loudspeaker was positioned in each spatial quadrant, on either side of the midline or the interaural axis, respectively. For four 8-speaker arrays, two loudspeakers were positioned in each quadrant, one close to the midline and the second separated from the first by 15 degrees, 30 degrees, 45 degrees, or 60 degrees. Three different 300-ms stimuli were localized: two one-third-octave noise bands, centered at 0.5 and 4 kHz, and broadband noise. The stimulus level (75 dB SPL) was well above hearing threshold for all subjects tested. Over the age range studied, percent-correct sound-source identification judgments decreased by 12%-15%. Performance decrements were apparent as early as the third decade of life. Broadband noise was easiest to localize (both binaural and spectral cues were available), and the 0.5-kHz noise band, the most difficult to localize (primarily interaural temporal difference cue available). Accuracy was relatively higher in front of than behind the head, and errors were largely front/back mirror image reversals. A left-sided superiority was evident until the fifth decade of life. The results support the conclusions that the processing of spectral information becomes progressively less efficient with aging, and is generally worse for sources on the right side of space.  相似文献   

3.
Narrow-band sound localization related to external ear acoustics.   总被引:3,自引:0,他引:3  
Human subjects localized brief 1/6-oct bandpassed noise bursts that were centered at 6, 8, 10, and 12 kHz. All testing was done under binaural conditions. The horizontal component of subjects' responses was accurate, comparable to that for broadband localization, but the vertical and front/back components exhibited systematic errors. Specifically, responses tended to cluster within restricted ranges that were specific for each center frequency. The directional transfer functions of the subjects' external ears were measured for 360 horizontal and vertical locations. The spectra of the sounds that were present in the subjects' ear canals, the "proximal stimulus" spectra, were computed by combining the spectra of the narrow-band sound sources with the directional transfer functions for particular stimulus locations. Subjects consistently localized sounds to regions within which the associated directional transfer function correlated most closely with the proximal stimulus spectrum. A quantitative model was constructed that successfully predicted subjects' responses based on interaural level difference and spectral cues. A test of the model, using techniques adapted from signal detection theory, indicated that subjects tend to use interaural level difference and spectral shape cues independently, limited only by a slight spatial correlation of the two cues. A testing procedure is described that provides a quantitative comparison of various predictive models of sound localization.  相似文献   

4.
For human listeners, cues for vertical-plane localization are provided by direction-dependent pinna filtering. This study quantified listeners' weighting of the spectral cues from each ear as a function of stimulus lateral angle, interaural time difference (ITD), and interaural level difference (ILD). Subjects indicated the apparent position of headphone-presented noise bursts synthesized in virtual auditory space. The synthesis filters for the two ears either corresponded to the same location or to two different locations separated vertically by 20 deg. Weighting of each ear's spectral information was determined by a multiple regression between the elevations to which each ear's spectrum corresponded and the vertical component of listeners' responses. The apparent horizontal source location was controlled either by choosing synthesis filters corresponding to locations on or 30 deg left or right of the median plane or by attenuating or delaying the signal at one ear. For broadband stimuli, spectral weighting and apparent lateral angle were determined primarily by ITD. Only for high-pass stimuli were weighting and lateral angle determined primarily by ILD. The results suggest that the weighting of monaural spectral cues and the perceived lateral angle of a sound source depend similarly on ITD, ILD, and stimulus spectral range.  相似文献   

5.
The influence of pinnae-based spectral cues on sound localization   总被引:1,自引:0,他引:1  
The role of pinnae-based spectral cues was investigated by requiring listeners to locate sound, binaurally, in the horizontal plane with and without partial occlusion of their external ears. The main finding was that the high frequencies were necessary for optimal performance. When the stimulus contained the higher audio frequencies, e.g., broadband and 4.0-kHz high-pass noise, localization accuracy was significantly superior to that recorded for stimuli consisting only of the lower frequencies (4.0- and 1.0-kHz low-pass noise). This finding was attributed to the influence of the spectral cues furnished by the pinnae, for when the stimulus composition included high frequencies, pinnae occlusion resulted in a marked decline in localization accuracy. Numerous front-rear reversals occurred. Moreover, the ability to distinguish among sounds originating within the same quadrant also suffered. Performance proficiency for the low-pass stimuli was not further degraded under conditions of pinnae occlusion. In locating the 4.0-kHz high-pass noise when both, neither, or only one ear was occluded, the data demonstrated unequivocally that the pinna-based cues of the "near" ear contributed powerfully toward localization accuracy.  相似文献   

6.
Four adult bilateral cochlear implant users, with good open-set sentence recognition, were tested with three different sound coding strategies for binaural speech unmasking and their ability to localize 100 and 500 Hz click trains in noise. Two of the strategies tested were envelope-based strategies that are clinically widely used. The third was a research strategy that additionally preserved fine-timing cues at low frequencies. Speech reception thresholds were determined in diotic noise for diotic and interaurally time-delayed speech using direct audio input to a bilateral research processor. Localization in noise was assessed in the free field. Overall results, for both speech and localization tests, were similar with all three strategies. None provided a binaural speech unmasking advantage due to the application of 700 micros interaural time delay to the speech signal, and localization results showed similar response patterns across strategies that were well accounted for by the use of broadband interaural level cues. The data from both experiments combined indicate that, in contrast to normal hearing, timing cues available from natural head-width delays do not offer binaural advantages with present methods of electrical stimulation, even when fine-timing cues are explicitly coded.  相似文献   

7.
Passive sound-localization acuity and its relationship to vision were determined for the echolocating Jamaican fruit bat (Artibeus jamaicensis). A conditioned avoidance procedure was used in which the animals drank fruit juice from a spout in the presence of sounds from their right, but suppressed their behavior, breaking contact with the spout, whenever a sound came from their left, thereby avoiding a mild shock. The mean minimum audible angle for three bats for a 100-ms noise burst was 10 degrees-marginally superior to the 11.6 degrees threshold for Egyptian fruit bats and the 14 degrees threshold for big brown bats. Jamaican fruit bats were also able to localize both low- and high-frequency pure tones, indicating that they can use both binaural phase- and intensity-difference cues to locus. Indeed, their ability to use the binaural phase cue extends up to 6.3 kHz, the highest frequency so far for a mammal. The width of their field of best vision, defined anatomically as the width of the retinal area containing ganglion-cell densities at least 75% of maximum, is 34 degrees. This value is consistent with the previously established relationship between vision and hearing indicating that, even in echolocating bats, the primary function of passive sound localization is to direct the eyes to sound sources.  相似文献   

8.
The cortical mechanisms of perceptual segregation of concurrent sound sources were examined, based on binaural detection of interaural timing differences. Auditory event-related potentials were measured from 11 healthy subjects. Binaural stimuli were created by introducing a dichotic delay of 500-ms duration to a narrow frequency region within a broadband noise, and resulted in a perception of a centrally located noise and a right-lateralized pitch (dichotic pitch). In separate listening conditions, subjects actively discriminated and responded to randomly interleaved binaural and control stimuli, or ignored random stimuli while watching silent cartoons. In a third listening condition subjects ignored stimuli presented in homogenous blocks. For all listening conditions, the dichotic pitch stimulus elicited an object-related negativity (ORN) at a latency of about 150-250 ms after stimulus onset. When subjects were required to actively respond to stimuli, the ORN was followed by a P400 wave with a latency of about 320-420 ms. These results support and extend a two-stage model of auditory scene analysis in which acoustic streams are automatically parsed into component sound sources based on source-relevant cues, followed by a controlled process involving identification and generation of a behavioral response.  相似文献   

9.
This paper evaluates the influence of three multimicrophone noise reduction algorithms on the ability to localize sound sources. Two recently developed noise reduction techniques for binaural hearing aids were evaluated, namely, the binaural multichannel Wiener filter (MWF) and the binaural multichannel Wiener filter with partial noise estimate (MWF-N), together with a dual-monaural adaptive directional microphone (ADM), which is a widely used noise reduction approach in commercial hearing aids. The influence of the different algorithms on perceived sound source localization and their noise reduction performance was evaluated. It is shown that noise reduction algorithms can have a large influence on localization and that (a) the ADM only preserves localization in the forward direction over azimuths where limited or no noise reduction is obtained; (b) the MWF preserves localization of the target speech component but may distort localization of the noise component. The latter is dependent on signal-to-noise ratio and masking effects; (c) the MWF-N enables correct localization of both the speech and the noise components; (d) the statistical Wiener filter approach introduces a better combination of sound source localization and noise reduction performance than the ADM approach.  相似文献   

10.
A human psychoacoustical experiment is described that investigates the role of the monaural and interaural spectral cues in human sound localization. In particular, it focuses on the relative contribution of the monaural versus the interaural spectral cues towards resolving directions within a cone of confusion (i.e., directions with similar interaural time and level difference cues) in the auditory localization process. Broadband stimuli were presented in virtual space from 76 roughly equidistant locations around the listener. In the experimental conditions, a "false" flat spectrum was presented at the left eardrum. The sound spectrum at the right eardrum was then adjusted so that either the true right monaural spectrum or the true interaural spectrum was preserved. In both cases, the overall interaural time difference and overall interaural level difference were maintained at their natural values. With these virtual sound stimuli, the sound localization performance of four human subjects was examined. The localization performance results indicate that neither the preserved interaural spectral difference cue nor the preserved right monaural spectral cue was sufficient to maintain accurate elevation judgments in the presence of a flat monaural spectrum at the left eardrum. An explanation for the localization results is given in terms of the relative spectral information available for resolving directions within a cone of confusion.  相似文献   

11.
High-frequency spectral notches are important cues for sound localization. Our ability to detect them must depend on their representation as auditory nerve (AN) rate profiles. Because of the low threshold and the narrow dynamic range of most AN fibers, these rate profiles deteriorate at high levels. The system may compensate by using onset rate profiles whose dynamic range is wider, or by using low-spontaneous-rate fibers, whose threshold is higher. To test these hypotheses, the threshold notch depth necessary to discriminate between a flat spectrum broadband noise and a similar noise with a spectral notch centered at 8 kHz was measured at levels from 32 to 100 dB SPL. The importance of the onset rate-profile representation of the notch was estimated by varying the stimulus duration and its rise time. For a large proportion of listeners, threshold notch depth varied nonmonotonically with level, increasing for levels up to 70-80 dB SPL and decreasing thereafter. The nonmonotonic aspect of the function was independent of notch bandwidth and stimulus duration. Thresholds were independent of stimulus rise time but increased for the shorter noise bursts. Results are discussed in terms of the ability of the AN to convey spectral notch information at different levels.  相似文献   

12.

Background  

In the field of auditory neuroscience, much research has focused on the neural processes underlying human sound localization. A recent magnetoencephalography (MEG) study investigated localization-related brain activity by measuring the N1m event-related response originating in the auditory cortex. It was found that the dynamic range of the right-hemispheric N1m response, defined as the mean difference in response magnitude between contralateral and ipsilateral stimulation, reflects cortical activity related to the discrimination of horizontal sound direction. Interestingly, the results also suggested that the presence of realistic spectral information within horizontally located spatial sounds resulted in a larger right-hemispheric N1m dynamic range. Spectral cues being predominant at high frequencies, the present study further investigated the issue by removing frequencies from the spatial stimuli with low-pass filtering. This resulted in a stepwise elimination of direction-specific spectral information. Interaural time and level differences were kept constant. The original, unfiltered stimuli were broadband noise signals presented from five frontal horizontal directions and binaurally recorded for eight human subjects with miniature microphones placed in each subject's ear canals. Stimuli were presented to the subjects during MEG registration and in a behavioral listening experiment.  相似文献   

13.
Sound localization cues generally include interaural time difference, interaural intensity difference, and spectral cues. The purpose of this study is to investigate the important spectral cues involved in so-called head related transfer functions (HRTFs) using a combination of HRTF analyses and a virtual sound localization (VSL) experiment. Previous psychoacoustical and physiological studies have both suggested the existence of spectral modulation frequency (SMF) channels for analyzing spectral information (e.g., the spectral cues coded in HRTFs). SMFs are in a domain related to the Fourier transform of HRTFs. The relationship between various SMF regions and sound localization was tested here by filtering or enhancing HRTFs in the SMF domain under a series of conditions using a VSL experiment. Present results revealed that azimuth localization was not significantly affected by HRTF manipulation. Applying notch filters between 0.1 and 0.4 cyclesoctave or between 0.35 and 0.65 cyclesoctave resulted in significantly less accurate elevation responses at low elevations, while spectral enhancement in these two SMF regions did not produce a significant change in sound localization. Likewise, low-pass filtering at 2 cyclesoctave did not significantly influence localization accuracy, suggesting that the major cues for sound localization are in the SMF region below 2 cyclesoctave.  相似文献   

14.
The effect of onset interaural time differences (ITDs) on lateralization and detection was investigated for broadband pulse trains 250 ms long with a binaural fundamental frequency of 250 Hz. Within each train, ITDs of successive binaural pulse pairs alternated between two of three values (0 micros, 500 micros left-leading, and 500 micros right-leading) or were invariant. For the alternating conditions, the experimental manipulation was the choice of which of two ITDs was presented first (i.e., at stimulus onset). Lateralization, which was estimated using a broadband noise pointer with a listener adjustable interaural delay, was determined largely by the onset ITD. However, detection thresholds for the signals in left-leading or diotic continuous broadband noise were not affected by where the signals were lateralized. A quantitative analysis suggested that binaural masked thresholds for the pulse trains were well accounted for by the level and phase of harmonic components at 500 and 750 Hz. Detection thresholds obtained for brief stimuli (two binaural pulse or noise burst pairs) were also independent of which of two ITDs was presented first. The control of lateralization by onset cues appears to be based on mechanisms not essential for binaural detection.  相似文献   

15.
A computational model of auditory localization resulting in performance similar to humans is reported. The model incorporates both the monaural and binaural cues available to a human for sound localization. Essential elements used in the simulation of the processes of auditory cue generation and encoding by the nervous system include measured head-related transfer functions (HRTFs), minimum audible field (MAF), and the Patterson-Holdsworth cochlear model. A two-layer feed-forward back-propagation artificial neural network (ANN) was trained to transform the localization cues to a two-dimensional map that gives the direction of the sound source. The model results were compared with (i) the localization performance of the human listener who provided the HRTFs for the model and (ii) the localization performance of a group of 19 other human listeners. The localization accuracy and front-back confusion error rates exhibited by the model were similar to both the single listener and the group results. This suggests that the simulation of the cue generation and extraction processes as well as the model parameters were reasonable approximations to the overall biological processes. The amplitude resolution of the monaural spectral cues was varied and the influence on the model's performance was determined. The model with 128 cochlear channels required an amplitude resolution of approximately 20 discrete levels for encoding the spectral cue to deliver similar localization performance to the group of human listeners.  相似文献   

16.
Eight listeners were required to locate a train of 4.5-kHz high-pass noise bursts emanating from loudspeakers positioned +/- 30, +/- 20, +/- 10, and 0 deg re: interaural axis. The vertical array of loudspeakers was placed at 45, 90, and 135 deg left of midline. The various experimental conditions incorporated binaural and monaural listening with the latter utilizing the ear nearest or ear farthest from the sound source. While performance excelled when listening with only the near ear, the contribution of the far ear was statistically significant when compared to localization performance when both ears were occluded. Based on head related transfer functions for stimuli whose bandwidth was 1.0 kHz, four spectral cues were selected as candidates for influencing location judgments. Two of them associated relative changes in energy across center frequencies (CFs) with vertical source positions. The other two associated an absolute minimum (maximum) energy for specific CFs with a vertical source position. All but one cue when measured for the near ear could account for localization proficiency. On the other hand, when listening with the far ear, maximum energy at a specific CF outperformed the remaining cues in accounting for localization proficiency.  相似文献   

17.
Localization dominance is an aspect of the precedence effect (PE) in which the leading source dominates the perceived location of a simulated echo (lagging source). It is known to be robust in the horizontal/azimuthal dimension, where binaural cues dominate localization. However, little is known about localization dominance in conditions that minimize binaural cues, and most models of precedence treat the phenomena as "belonging" to the binaural system. Here, localization dominance in the median-sagittal plane was studied where binaural cues are greatly reduced, and monaural spectral/level cues are thought to be the primary cues used for localization. Lead-lag pairs of noise bursts were presented from locations spaced in 15 degrees increments in the frontal, median-sagittal plane, with a 2-ms delay in their onsets, for source durations of 1, 10, 25, and 50-ms. Intermixed with these trials were single-speaker trials, in which lead and lag were summed and presented from one speaker. Listeners identified the speaker that was nearest to the perceived source location. With single-speaker stimuli, localization improves as signal duration is increased. Furthermore, evidence of elevation compression was found with a dependence on duration. With lead-lag pairs, localization dominance occurs in the median plane, and becomes more robust with increased signal duration. These results suggest that accurate localization of a co-located lead-lag pair is necessary for localization dominance to occur when the lag is spatially separated from the lead.  相似文献   

18.
Two sound localization algorithms based on the head-related transfer function were developed. Each of them uses the interaural time delay, interaural level difference, and monaural spectral cues to estimate the location of a sound source. Given that most localization algorithms will be required to function in background noise, the localization performance of one of the algorithms was tested at signal-to-noise ratios (SNRs) from 40 to -40 dB. Stimuli included ten real-world, broadband sounds located at 5 degrees intervals in azimuth and at 0 degrees elevation. Both two- and four-microphone versions of the algorithm were implemented to localize sounds to 5 degrees precision. The two-microphone version of the algorithm exhibited less than 2 degrees mean localization error at SNRs of 20 dB and greater, and the four-microphone version committed approximately 1 degrees mean error at SNRs of 10 dB or greater. Potential enhancements and applications of the algorithm are discussed.  相似文献   

19.
Binaural disparities are the primary acoustic cues employed in sound localization tasks. However, the degree of binaural correlation in a sound serves as a complementary cue for detecting competing sound sources [J. F. Culling, H. S. Colburn, and M. Spurchise, "Interaural correlation sensitivity," J. Acoust. Soc. Am. 110(2), 1020-1029 (2001) and L. R. Bernstein and C. Trahiotis, "On the use of the normalized correlation as an index of interaural envelope correlation," J. Acoust. Soc. Am. 100, 1754-1763 (1996)]. Here a random chord stereogram (RCS) sound is developed that produces a salient pop-out illusion of a slowly varying ripple sound [T. Chi et al., "Spectro-temporal modulation transfer functions and speech intelligibility," J. Acoust. Soc. Am. 106(5), 2719-2732 (1999)], even though the left and right ear sounds alone consist of noise-like random modulations. The quality and resolution of this percept is systematically controlled by adjusting the spectrotemporal correlation pattern between the left and right sounds. The prominence and limited time-frequency resolution for resolving the RCS suggests that envelope correlations are a dominant binaural cue for grouping acoustic objects.  相似文献   

20.
Localization responses to a broadband noise signal presented against a broadband noise masker were obtained from 12-month-old infants and adults. Two loudspeakers, one to the left and one to the right of the listener, continuously presented identical broadband maskers. On a trial, a broadband signal was added to one of the loudspeakers. Subjects were required to identify the loudspeaker producing the signal. Noise signals were either coherent (from the same noise generator) or incoherent (from an independent noise generator). Both infants and adults found it easier to locate the incoherent signals even when the two types of signals were adjusted to produce equal increments in power. Since monaural performance, after this adjustment, should be equivalent for the two cases, superior performance for incoherent signals implies that binaural processing is involved. The same result was observed in control experiments in which coherent and incoherent signals were presented over earphones to adults. These results suggest that the mechanisms responsible for binaural unmasking are operative by 12 months of age.  相似文献   

设为首页 | 免责声明 | 关于勤云 | 加入收藏

Copyright©北京勤云科技发展有限公司  京ICP备09084417号