Human sensitivity to differences in the rate of auditory cue change.
Ontology highlight
ABSTRACT: Measurement of sensitivity to differences in the rate of change of auditory signal parameters is complicated by confounds among duration, extent, and velocity of the changing signal. Dooley and Moore [(1988) J. Acoust. Soc. Am. 84(4), 1332-1337] proposed a method for measuring sensitivity to rate of change using a duration discrimination task. They reported improved duration discrimination when an additional intensity or frequency change cue was present. The current experiments were an attempt to use this method to measure sensitivity to the rate of change in intensity and spatial position. Experiment 1 investigated whether duration discrimination was enhanced when additional cues of rate of intensity change, rate of spatial position change, or both were provided. Experiment 2 determined whether participant listening experience or the testing environment influenced duration discrimination task performance. Experiment 3 assessed whether duration discrimination could be used to measure sensitivity to rates of changes in intensity and spatial position for stimuli with lower rates of change, as well as emphasizing the constancy of the velocity cue. Results of these experiments showed that duration discrimination was impaired rather than enhanced by the additional velocity cues. The findings are discussed in terms of the demands of listening to concurrent changes along multiple auditory dimensions.
Project description:For amplitude-modulated sound, the envelope interaural time difference (ITDENV) is a potential cue for sound-source location. ITDENV is encoded in the lateral superior olive (LSO) of the auditory brainstem, by excitatory-inhibitory (EI) neurons receiving ipsilateral excitation and contralateral inhibition. Between human listeners, sensitivity to ITDENV varies considerably, but ultimately decreases with increasing stimulus carrier frequency, and decreases more strongly with increasing modulation rate. Mechanisms underlying the variation in behavioral sensitivity remain unclear. Here, with increasing carrier frequency (4-10 kHz), as we phenomenologically model the associated decrease in ITDENV sensitivity using arbitrarily fewer neurons consistent across populations, we computationally model the variable sensitivity across human listeners and modulation rates (32-800 Hz) as the decreasing range of membrane frequency responses in LSO neurons. Transposed tones stimulate a bilateral auditory-periphery model, driving model EI neurons where electrical membrane impedance filters the frequency content of inputs driven by amplitude-modulated sound, evoking modulation filtering. Calculated from Fisher information in spike-rate functions of ITDENV, for model EI neuronal populations distinctly reflecting the LSO range in membrane frequency responses, just-noticeable differences in ITDENV collectively reproduce the largest variation in ITDENV sensitivity across human listeners. These slow to fast model populations each generally match the best human ITDENV sensitivity at a progressively higher modulation rate, by membrane-filtering and spike-generation properties producing realistically less than Poisson variance. Non-resonant model EI neurons are also sensitive to interaural intensity differences. With peripheral filters centered between carrier frequency and modulation sideband, fast resonant model EI neurons extend ITDENV sensitivity above 500-Hz modulation.
Project description:Electrophysiological evidence suggested primarily the involvement of the middle temporal (MT) area in depth cue integration in macaques, as opposed to human imaging data pinpointing area V3B/kinetic occipital area (V3B/KO). To clarify this conundrum, we decoded monkey functional MRI (fMRI) responses evoked by stimuli signaling near or far depths defined by binocular disparity, relative motion, and their combination, and we compared results with those from an identical experiment previously performed in humans. Responses in macaque area MT are more discriminable when two cues concurrently signal depth, and information provided by one cue is diagnostic of depth indicated by the other. This suggests that monkey area MT computes fusion of disparity and motion depth signals, exactly as shown for human area V3B/KO. Hence, these data reconcile previously reported discrepancies between depth processing in human and monkey by showing the involvement of the dorsal stream in depth cue integration using the same technique, despite the engagement of different regions.
Project description:The present study demonstrates, for the first time, a specific enhancement of auditory spatial cue discrimination due to eye gaze. Whereas the region of sharpest visual acuity, called the fovea, can be directed at will by moving one's eyes, auditory spatial information is derived primarily from head-related acoustic cues. Past auditory studies have found better discrimination in front of the head [1-3] but have not manipulated subjects' gaze, thus overlooking potential oculomotor influences. Electrophysiological studies have shown that the inferior colliculus, a critical auditory midbrain nucleus, shows visual and oculomotor responses [4-6] and modulations of auditory activity [7-9], and that auditory neurons in the superior colliculus show shifting receptive fields [10-13]. How the auditory system leverages this crossmodal information at the behavioral level remains unknown. Here we directed subjects' gaze (with an eccentric dot) or auditory attention (with lateralized noise) while they performed an auditory spatial cue discrimination task. We found that directing gaze toward a sound significantly enhances discrimination of both interaural level and time differences, whereas directing auditory spatial attention does not. These results show that oculomotor information variably enhances auditory spatial resolution even when the head remains stationary, revealing a distinct behavioral benefit possibly arising from auditory-oculomotor interactions at an earlier level of processing than previously demonstrated.
Project description:In speech, listeners extract continuously-varying spectrotemporal cues from the acoustic signal to perceive discrete phonetic categories. Spectral cues are spatially encoded in the amplitude of responses in phonetically-tuned neural populations in auditory cortex. It remains unknown whether similar neurophysiological mechanisms encode temporal cues like voice-onset time (VOT), which distinguishes sounds like /b/ and/p/. We used direct brain recordings in humans to investigate the neural encoding of temporal speech cues with a VOT continuum from /ba/ to /pa/. We found that distinct neural populations respond preferentially to VOTs from one phonetic category, and are also sensitive to sub-phonetic VOT differences within a population's preferred category. In a simple neural network model, simulated populations tuned to detect either temporal gaps or coincidences between spectral cues captured encoding patterns observed in real neural data. These results demonstrate that a spatial/amplitude neural code underlies the cortical representation of both spectral and temporal speech cues.
Project description:Unlike vision, the mechanisms underlying auditory motion perception are poorly understood. Here we describe an auditory motion illusion revealing a novel cue to auditory speed perception: the temporal frequency of amplitude modulation (AM-frequency), typical for rattling sounds. Naturally, corrugated objects sliding across each other generate rattling sounds whose AM-frequency tends to directly correlate with speed. We found that AM-frequency modulates auditory speed perception in a highly systematic fashion: moving sounds with higher AM-frequency are perceived as moving faster than sounds with lower AM-frequency. Even more interestingly, sounds with higher AM-frequency also induce stronger motion aftereffects. This reveals the existence of specialized neural mechanisms for auditory motion perception, which are sensitive to AM-frequency. Thus, in spatial hearing, the brain successfully capitalizes on the AM-frequency of rattling sounds to estimate the speed of moving objects. This tightly parallels previous findings in motion vision, where spatio-temporal frequency of moving displays systematically affects both speed perception and the magnitude of the motion aftereffects. Such an analogy with vision suggests that motion detection may rely on canonical computations, with similar neural mechanisms shared across the different modalities.
Project description:Environmental temperature affects physiological functions, representing a barrier for the range expansions of ectothermic species. To understand the link between thermal physiology and biogeography, a key question is whether among-species thermal sensitivity of metabolic rates is mechanistically constrained or buffered through physiological remodeling over evolutionary time. The former conception, the Universal Temperature Dependence hypothesis, predicts similar among- and within-species thermal sensitivity. The latter conception, the Metabolic Cold Adaptation hypothesis, predicts lower among-species thermal sensitivity than within-species sensitivity. Previous studies that tested these hypotheses for fishes overwhelmingly investigated teleosts with elasmobranchs understudied. Here, we show that among-species thermal sensitivity of resting metabolic rates is lower than within-species sensitivity in teleosts but not in elasmobranchs. Further, species richness declines with latitude more rapidly in elasmobranchs than in teleosts. Metabolic Cold Adaptation exhibited by teleosts might underpin their high diversity at high latitudes, whereas the inflexible thermal sensitivity approximated by Universal Temperature Dependence of elasmobranchs might explain their low diversity at high latitudes.
Project description:BACKGROUND: We used PET to study cortical activation during auditory stimulation and found sex differences in the human primary auditory cortex (PAC). Regional cerebral blood flow (rCBF) was measured in 10 male and 10 female volunteers while listening to sounds (music or white noise) and during a baseline (no auditory stimulation). RESULTS AND DISCUSSION: We found a sex difference in activation of the left and right PAC when comparing music to noise. The PAC was more activated by music than by noise in both men and women. But this difference between the two stimuli was significantly higher in men than in women. To investigate whether this difference could be attributed to either music or noise, we compared both stimuli with the baseline and revealed that noise gave a significantly higher activation in the female PAC than in the male PAC. Moreover, the male group showed a deactivation in the right prefrontal cortex when comparing noise to the baseline, which was not present in the female group. Interestingly, the auditory and prefrontal regions are anatomically and functionally linked and the prefrontal cortex is known to be engaged in auditory tasks that involve sustained or selective auditory attention. Thus we hypothesize that differences in attention result in a different deactivation of the right prefrontal cortex, which in turn modulates the activation of the PAC and thus explains the sex differences found in the activation of the PAC. CONCLUSION: Our results suggest that sex is an important factor in auditory brain studies.
Project description:The effect of stimulus modulation rate on the underlying neural activity in human auditory cortex is not clear. Human studies (using both invasive and noninvasive techniques) have demonstrated that at the population level, auditory cortex follows stimulus envelope. Here we examined the effect of stimulus modulation rate by using a rare opportunity to record both spiking activity and local field potentials (LFP) in auditory cortex of patients during repeated presentations of an audio-visual movie clip presented at normal, double, and quadruple speeds. Mean firing rate during evoked activity remained the same across speeds and the temporal response profile of firing rate modulations at increased stimulus speeds was a linearly scaled version of the response during slower speeds. Additionally, stimulus induced power modulation of local field potentials in the high gamma band (64-128 Hz) exhibited similar temporal scaling as the neuronal firing rate modulations. Our data confirm and extend previous studies in humans and anesthetized animals, supporting a model in which both firing rate, and high-gamma LFP power modulations in auditory cortex follow the temporal envelope of the stimulus across different modulation rates.
Project description:Listeners assign different weights to spectral dynamics, such as formant rise time (FRT), and temporal dynamics, such as amplitude rise time (ART), during phonetic judgments. We examined the neurophysiological basis of FRT and ART weighting in the /ba/-/wa/ contrast. Electroencephalography was recorded for thirteen adult English speakers during a mismatch negativity (MMN) design using synthetic stimuli: a /ba/ with /ba/-like FRT and ART; a /wa/ with /wa/-like FRT and ART; and a /ba/(wa) with /ba/-like FRT and /wa/-like ART. We hypothesized that because of stronger reliance on FRT, subjects would encode a stronger memory trace and exhibit larger MMN during the FRT than the ART contrast. Results supported this hypothesis. The effect was most robust in the later portion of MMN. Findings suggest that MMN is generated by multiple sources, differentially reflecting acoustic change detection (earlier MMN, bottom-up process) and perceptual weighting of ART and FRT (later MMN, top-down process).
Project description:In this work, a binaural model resembling the human auditory system was built using a pair of three-dimensional (3D)-printed ears to localize a sound source in both vertical and horizontal directions. An analysis on the proposed model was firstly conducted to study the correlations between the spatial auditory cues and the 3D polar coordinate of the source. Apart from the estimation techniques via interaural and spectral cues, the property from the combined direct and reverberant energy decay curve is also introduced as part of the localization strategy. The preliminary analysis reveals that the latter provides a much more accurate distance estimation when compared to approximations via sound pressure level approach, but is alone not sufficient to disambiguate the front-rear confusions. For vertical localization, it is also shown that the elevation angle can be robustly encoded through the spectral notches. By analysing the strengths and shortcomings of each estimation method, a new algorithm is formulated to localize the sound source which is also further improved by cross-correlating the interaural and spectral cues. The proposed technique has been validated via a series of experiments where the sound source was randomly placed at 30 different locations in an outdoor environment up to a distance of 19 m. Based on the experimental and numerical evaluations, the localization performance has been significantly improved with an average error of 0.5 m from the distance estimation and a considerable reduction of total ambiguous points to 3.3%.