Link to this chapter Copy to clipboard
Cite this chapter Copy to clipboard
Embed this chapter on your site Copy to clipboard
Embed this code snippet in the HTML of your website to show this chapter
Open access peer-reviewed chapter
By Deborah Hall, Cornelis Lanting, Douglas Hartley
Reviewed: January 23rd 2014Published: May 31st 2014
During the last two decades or so, auditory neuroscience has made significant progress in understanding the functional organization of the auditory system in both normal-hearing listeners, patients with sensorineural hearing impairments and/or tinnitus, and following cochlear implantation. Modern brain imaging techniques have made an enormous contribution to that progress by enabling the in vivo study of human central auditory function. Since its introduction, magnetic resonance imaging (MRI) has evolved to become one of the most influential methods for investigating non-invasive human brain structure and function. Functional MRI (fMRI) has become the tool of choice for addressing many research questions concerning central auditory plasticity. First, fMRI has good spatial sensitivity and specificity and so can be used to map auditory responses with reasonably fine topographic detail (Logothetis et al., 2001). Second, fMRI is suitable for research use with children, as well as adults. Third, since multiple observations can be made on the same individual, fMRI permits the investigation of longer-term dynamic processes, such as functional plasticity after disease, damage, retraining or therapy. Finally, the need for averaging data across individuals is reduced, further improving the accuracy with which activations can be mapped onto subject-specific brain structure (e.g. Talavage et al., 2004 ). Despite these advantages, there are significant issues using this technique in certain clinical populations, such as individuals with a cochlear implant (CI). Subsequently, to investigate the effects of CIs on brain reorganization, it is necessary to consider alternative functional imaging methods, such as electroencephalography (e.g. Debener et al., 2008). Moreover, the high level of the scanner acoustic noise can have detrimental effects on central auditory function, patient anxiety and verbal communication between patient and scanner operator (Mechefske et al., 2002). In terms of central auditory function, the scanner acoustic noise can induce increases in the fMRI response of the auditory cortex as well as interfere with the perception of externally presented sound stimuli (Edmister et al., 1999; Hall et al., 1999).
Although- functional imaging of the central auditory pathway has not been widely used in Ear, Nose and Throat or Audiology clinics, there is a growing literature on its potential application. Using insights gained from functional imaging research with normal-hearing subjects, we can now transfer the most efficient protocols to study patients with hearing problems. For example, a choice of methods is available for noise reduction and a range of scanning parameters have been optimised to enhance the detection of brain activity.
This chapter discusses some of the theoretical and technical challenges of auditory fMRI and several of the major themes within this field, notably cortical reorganization underpinning tinnitus and deafness. Also we will consider the effect of CIs on the auditory brain, the difficulties associated with the use of conventional imaging methods in CI populations, and potential application of novel techniques, such as near infrared spectroscopy. Sections not only highlight evidence for neuroplasticity within the central auditory system, but also review what is currently known about the involvement and integration of information from other sensory modalities.
Rather than directly measuring the stimulus- and task-related electrical activity of neurons, fMRI measures indirect changes in the magnetic properties of the blood. Most common methods make use of the blood-oxygen-level-dependent (BOLD) contrast. This technique is based on the increase in signal intensity caused by an increase in oxygen concentration of blood (Ogawa et al., 1990). The functional image represents the spatial distribution of blood oxygenation levels in the brain and the small fluctuations in these levels over time are correlated with the stimulus input or cognitive task.
To detect the magnetic properties of brain tissue, MR scanners operate using three different types of electromagnetic fields: a very high static field generated by a superconducting magnet, time-varying gradient magnetic fields and pulsed radiofrequency (RF) fields. The latter two fields are much weaker than the first, but all pose a number of unique and considerable technical challenges for conducting auditory fMRI research within this hostile environment.
In the first place, the static and time-varying magnetic fields preclude the use of many types of electronic sound presentation equipment, as well as preventing the safe scanning of patients who are wearing listening devices such as hearing aids or implants. Despite the restriction on the materials that can be used in a scanner, a number of different MR-compatible active headphone driving units have been produced for sound delivery. Several audio systems use electrostatic transducers to ensure high performance. Electrostatic headphones generate sound using a conductive diaphragm placed next to a fixed conducting panel. A high voltage polarises the fixed panel and the audio signal passing through the diaphragm rapidly switches between a positive and a negative signal, attracting or repelling it to the fixed panel and thus vibrating the air. Induced currents in the electronics, caused directly by the time-varying gradient magnetic fields or the RF pulses, are an additional hazard to the electronic devices themselves. Some materials can also absorb the RF energy causing local tissue heating and even burns if in contact with soft tissue. Hence, the digital audio source, electronics and power supply that drive the sound system are housed outside the RF screened scanner room to avoid electromagnetic interference with MRI. The electrical signal is then transferred from the audio source to the headphones in the RF screened scanner room using either filters through a filter panel or fibre-optic cable through the waveguide. For the same safety reasons, there are restrictions on scanning people who have electronic listening devices. These include hearing aids, cochlear implants and brainstem implants. Official approval for the manufacture of implant devices requires rigorous testing for susceptibility to electromagnetic fields, radiated electromagnetic fields and electrical safety testing (including susceptibility to electrical discharge). Typically however standard listening devices do not meet the more stringent criteria for compatibility with MR scanning. For the patient, risks include movement of the device and localised heating of brain tissue, whereas, for the device, the electronic components may be damaged. Some implant designs have been proven to be MR compatible (e.g. Shellock et al., 1993), but they are not routinely supplied in clinical practice. Given that implanted devices cannot be removed without surgical intervention, clinical imaging research of implantees has generally used other brain imaging methods, namely positron emission tomography (Giraud et al., 2001).
The high levels of scanner noise generated by the flexing of the gradient coils in the static magnetic field can reduce audibility and increase listening effort. Scanner noise increases nonlinearly with static magnetic field strength. For example, comparing scanners from 1 to 3 Tesla indicates a difference in sound level of about 10 dB SPL (Price et al., 2001). The sound level measured in the bore of the scanner is typically 99 dB SPL (98 dB(A)). Clearly, exposure to such an intense sound levels without protection is likely to cause a temporary threshold shift in hearing and tinnitus, and could be permanently damaging over a prolonged stay inside the bore of the scanner (Foster et al., 2000). The simplest way to treat the intense noise is to use ear protection in the form of ear defenders and/or ear plugs. Foam ear plugs can compromise the acoustic quality of the experimental sounds delivered to the subject and so circumaural headsets are preferable. Typically transducers are fitted into sound attenuating earmuffs to reduce the ambient noise level at the subject’s ears. Considerable attenuation can be achieved in this manner (e.g., 25 dB of attenuation at 600 Hz rising to 40 dB at 4 kHz, Hall et al., 2009). Even with hearing protection, whenever the scanner noise coincides with the presented sound stimulus it produces changes in task performance and increases the attentional demands of the listening task. The frequency range of the scanner noise overlaps with the range that is crucial for speech intelligibility and so speech experiments can be particularly compromised in the noisy environment. For example, Blackman and Hall (2011) reported that speech intelligibility (measured as the proportion of keywords reported correctly) dropped by about 50% for words that were masked by the scanner noise compared to those that were not masked by the noise.
The high levels of ambient scanner noise can interfere with the patterns of brain activity measured during fMRI. The noise not only activates parts of the auditory brain, but also interacts with the patterns of activity evoked by experimental stimuli. Two novel methods have been developed by the Nottingham group to deal with these problems; sparse temporal sampling (Hall et al., 1999) and active noise reduction (Hall et al., 2009). In sparse sampling, a clustered acquisition of brain slices is combined with a long and fixed (e.g. 10 s) inter-scan interval and the scan acquisition is temporally offset from the presentation of the stimulus of interest (typically by 8-10 s). Although the scanner noise resulting from the acquisition still evokes an auditory BOLD response, it does not extend across time to the subsequent scan. Work in the mid-90s demonstrated that when the onset of the scan acquisition is synchronized to the offset of an epoch of the stimulus presentation in a sparse sampling paradigm, the activation associated with the experimental sound can be effectively separated from the activation associated with the scanner sound (Hall et al., 1999). Furthermore, because the scanner sound is temporally offset, it does not produce acoustical masking and does not distract the listener. Sparse sampling is often the scanning protocol of choice for identifying auditory cortical evoked responses in the absence of scanner noise. Active noise reduction achieves significant attenuation of the acoustic energy received at the ears. The system reduces the perceived level of the background noise by producing an output (‘cancellation’) signal that destructively interferes with the scanner sound (i.e. the incident sound wave) at each ear. Precise control over the frequency and phase of acoustic signals presented to participants is achieved through a digital-signal processing system that delivers sound stimuli via a high-fidelity electrostatic headset. Blackman and Hall (2011) evaluated the benefit of active noise reduction by asking whether sound reduction provided sufficient gains in both audibility and sensitivity to sound-related activity to remove the need for sparse temporal sampling. While both sparse sampling and cancellation were found to improve listening quality, these factors did not provide equivalent benefits for auditory fMRI experiments. Cancellation offered small improvements for spatially specific analyses within Heschl’s gyrus and planum polare '(regions that are engaged in auditory perception and speech comprehension)' and for reducing some of the extra-auditory activity associated with the effort required for perceptual discrimination in a noisy environment. Nevertheless, sparse acquisition remained the more robust method for detecting auditory cortical activity by providing proper quiet intervals for stimulus presentation.
Brain imaging studies in humans are revealing parallels with the functional organization of the auditory brain discovered in electrophysiological studies in animals: sound-level dependencies as well as topographical mapping of stimulus frequencies are observed. This section provides a foundation to the remaining sections in this chapter by describing the BOLD effect and highlighting experiments that investigate the spatial specificity of this effect as a function of the external magnetic field strength and the parameters of the imaging sequence. We consequently describe fMRI experiments that focus on the functional organization of the auditory system. Specifically, we report an experiment on the topographical mapping of sound frequencies in human auditory cortex and discuss common findings and interpretations.
Research systems typically operate at 3 Tesla, but nowadays MRI systems that operate at magnetic field strengths exceeding this field strength are becoming more widely available. Nottingham University is the home of the first ultra-high field 7 Tesla MRI scanner for human brain imaging in the UK. The key factor for the development of such high field MRI systems is the improvement in intrinsic signal-to-noise ratio and contrast-to-noise ratio with increasing field strength. A simple motor task, for example, elicits a signal change of approximately 1%, 2.5% and 5% for field strengths of respectively 1.5 Tesla, 3 Tesla and 7 Tesla (Van der Zwaag et al., 2009). Figure 1 shows the results for one of our typical auditory tasks acquired at a magnetic field strength of 7 Tesla. In this short experiment, a sound was alternated with silence in a block design. The signal change was 5.2% indicating that the obtained percent signal change is comparable to that obtained in the above motor task. The resulting increased BOLD effect is of great benefit for fMRI basic research studies and also for clinically focussed studies. It can be exploited to improve the spatial resolution and reduces the number of trials required to demonstrate significant levels of activation. An additional effect is the increased spatial specificity of the BOLD effect with increasing field strength. At higher field strengths the T2*time of blood is relatively short causing its signal, originating mainly from the vasculature, to be attenuated relative to that of tissue at the echo times used for BOLD fMRI (Yacoub et al., 2001). The contribution of venous signal to the BOLD effect is therefore reduced at high field strengths, increasing the spatial specificity of the BOLD effect.
The spatial resolution of the BOLD effect is, compared to most MRI imaging techniques, quite poor. T1-weighted anatomical images can, for example, be easily obtained with an isotropic resolution of well below 1 mm. In contrast, the spatial resolution of the BOLD effect is mostly determined by physiological sources, in the sense that the vascular response to neural activity extends over several millimetres (Logothetis et al., 2001). The BOLD signal originating from small capillaries and surrounding areas is thought to provide good spatial localization for neuronal activation. However, at low and intermediate field strengths, the gradient echo (GE) BOLD effects from large draining veins often contaminate the signal, causing inaccurate localization of the activity and also cause a widening of the point spread function of the BOLD response. GE BOLD is sensitive to extravascular effects around the large vessels while spin-echo (SE) sequences are relatively insensitive to these effects (Duong et al., 2003; Weisskoff et al., 1994). The 180° rephasing pulse in an SE sequence compensates for the constant field heterogeneities to obtain an echo that is weighted in T2 and not in T2*. Spin-echo sequences are therefore more closely related to neuronal activity than GE sequences, although there is still a large contribution of intravascular signal arising from blood. The benefit of high fields (i.e. at field strengths of 3 Tesla and higher) compared to a lower field strength is the reduced T2 and T2* of blood relative to that of tissue, reducing the intravascular component even more (Yacoub et al., 2001). However, a major disadvantage with T2-weighted SE sequences is linked to long inter-scan intervals, resulting in potentially prohibitive acquisition times.
The point spread function describes the response of the MRI scanning to a point source within the brain. Measurements of the point spread function have revealed a full-width at half-maximum (FWHM) of around 2.3 mm tangential to cortical surface of the visual cortex using GE BOLD at 7 Tesla (Shmuel et al., 2007). The FWHM of the point spread function at 7 Tesla is therefore narrower than at 1.5 Tesla (reported at 3.5 mm, Engel et al., 1997) and at 3 Tesla (reported at 3.9 mm, Parkes et al., 2005). Recently, Polimeni et al. (2010) measured the spatial spread of the BOLD response tangential to the cortical surface as a function of cortical depth. Sampling near the pial surface provided the highest signal strength, but it also introduced the most spatial error due to extravascular BOLD effects. Avoiding the surface laminae improved spatial localization considerably, at the cost of sensitivity, implying that optimal spatial resolution in fMRI of the cortex can be achieved using anatomically informed spatial sampling to avoid large pial vessels.
When comparing the point spread function of GE and SE BOLD sequences there is a 13% reduction of the FWHM of the SE sequence, but at a cost of at least 3-fold reduction in contrast to noise ratio (Parkes et al., 2005). fMRI suffers from susceptibility artifacts that are caused by the proximity of air-cavities like the ear canal and nasal cavity causing signal loss and distortion of the images (Ojemann et al., 1997). Many strategies have been proposed to overcome these effects which include the usage of shorter echo times, parallel imaging methods and improved shimming methods. Since static susceptibility artifacts can –at least partly- be refocused using SE techniques, it is less affected by susceptibility artifacts than GE methods and may increase the sensitivity in affected areas (Goense et al., 2008) such as auditory regions in the temporal lobe.
In conclusion, although there is a trade-off between sensitivity and specificity, and the fact that there are still some technical considerations that need to be assessed, most recent imaging literature agree on the advantages of ultra-high field fMRI (i.e. ≥ 7 Tesla), that is, the increased signal-to-noise ratio. This evolution is of special importance for the clinical field in general and the field of auditory research specifically. The next section discusses the application of fMRI to study the functional organization of the auditory cortex.
Many studies have elucidated the functional specialization and hierarchical organization of multiple areas of the auditory cortex in non-human primates using invasive neurophysiological recordings which are obviously not applicable to human studies. A method to map frequency mapping using noninvasive BOLD fMRI in primates is in its early infancy (Petkov, personal communication). At present, there are many applications of BOLD fMRI in experiments on the human auditory system, ranging from the characterization of the effects of sound intensity (Bilecen et al., 2002; Hart et al., 2003), the effects of the side of presentation (Jäncke et al., 2002) to more complex phenomena such as the interaction of scanner noise and sound-evoked responses (Gaab et al., 2007; Hall et al. 1999; Talavage & Edmister 2004 ) and feature sensitivity and attention modulation (Paltoglou et al., 2011). Mapping one of the most fundamental properties of the auditory system, that of frequency coding, remains elusive and only a few studies have demonstrated a clear topographical arrangement of the response to sound frequency in the auditory pathway. This organization, referred to as ‘tonotopy’, mirrors the distribution of receptors in the cochlea, with neurons at the base of the cochlea preferentially responding to high frequencies and neurons at the apex responding best to low frequencies (Ehret, 1978). Tonotopy is conveyed up to the level of auditory cortex by fibres in the lemniscal pathway (Figure 3). In the auditory cortex of nonhuman primates, the primary core cortex, consisting of A1, R and RT, exhibit tonotopic gradients that are mirror symmetric to each other (Hackett et al., 2001; Kaas & Hackett 2000; Morel et al., 1993).
Tonotopic organization has been identified in human auditory cortex using a variety of imaging techniques including positron emission tomography (PET) and fMRI. The majority of early studies were based on just two different stimulus frequencies (Bilecen et al., 1998; Lauter et al., 1985; Lockwood et al. 1999; Talavage et al., 2000; Wessinger et al., 1997). From the results a general pattern emerges where high frequencies are represented at the medial end of auditory cortex and low frequencies in the anterior and lateral part of Heschl’s gyrus and surrounding areas, mostly as a single low-to-high frequency gradient. More recent studies used more frequencies allowing the identification of more fine-grained frequency gradients. However, the results and interpretations of these studies vary considerably. For example, Langers et al. (2007) reported a single high-to-low gradient extending from posterior medial to anterior lateral auditory areas, similar to earlier studies. Mapping of the left hemisphere at 7 Tesla using surface coils and sparse image acquisition revealed two mirror-symmetric tonotopic maps encompassing three linear gradients of frequency-sensitive BOLD responses in six subjects: (i) one connecting a high-frequency region in Heschl’s sulcus with a low-frequency region halfway Heschl’s gyrus and (ii) a second connected the same low-frequency region with a high-frequency region near the lateral junction of Heschl’s gyrus and the superior temporal gyrus (Formisano et al., 2003). These results suggest that human auditory cortex conforms to the primate model with two mirror-symmetric tonotopic maps in primary auditory core and belt areas that are joined at a common low-frequency boundary (Petkov et al., 2006). Talavage and colleagues reported three consistent gradients, none of which clearly followed the long (medial-to-lateral) axis of Heschl’s gyrus ( Talavage et al., 2004 ). Schönwiesner et al. (2002) argued that the observed frequency selectivity does not necessarily represent frequency gradients but instead represent different functional regions within the auditory cortex (Schönwiesner et al., 2002). Based on this assumption, in combination with the models by Kaas et al. (2000), three recent studies investigated tonotopic mapping and related their findings to anatomical models.
Humphries reported two major mirror-symmetric frequency gradients, oriented roughly perpendicular to Heschl’s gyrus, rather than parallel as previously assumed (Humphries et al., 2010). They found a single low-frequency area centered on Heschl’s gyrus, bordered anteriorly and posteriorly by two high-frequency areas (planum temporale). This seems very similar to previous findings (Woods et al., 2009). Based on the observed mirror-symmetric frequency gradients, both sets of authors proposed boundaries of regions in human auditory cortex homologous to A1 and R regions in non-human primate cortex. Figure 2 shows the schematic outline of auditory fields as based on electrophysiology in primates and the original group data of both Woods et al. (2009) and Humphries et al. (2010) indicating similar tonotopic maps and a close correspondence between these maps and proposed boundaries as compared to primate data.
A different approach was taken by Striem-Amit et al. (2011), who used phase-encoding (travelling wave) fMRI, a technique previously used to map the visual field onto the visual cortex (Engel et al., 1997; Warnking, 2002). They found up to six mirror-symmetric frequency gradients covering most of Heschl’s gyrus and extending to the superior temporal gyrus and superior temporal sulcus (Striem-Amit et al., 2011), beyond the areas previously reported by Woods et al. (2009) and Humphries et al. (2010). Such cochleotopic organization was identified in both left and right hemispheres. The maps of Striem-Amit et al. (2011) demonstrate that tonotopy persists well beyond the auditory core and belt.
The discovery of the tonotopic maps and proposed boundaries between auditory cortical fields might help our understanding of these fields and contributes in defining possible connectivity patterns between these fields. These advances ultimately help us understand the functional consequences of e.g., hearing loss and tinnitus which may affect the function or organization of the auditory cortex (Eggermont, 2006; Komiya & Eggermont, 2000). Both conditions are good examples of auditory disorders with an increasing prevalence due to an aging population.
Tinnitus is an auditory sensation without the presence of an external acoustic stimulus. In 6-20% of the adult population, tinnitus is a chronic condition and severely disrupts quality of life in 1-3%. Tinnitus is more prevalent in men than in women and its prevalence increases with advancing age (Axelsson & Ringdahl, 1989). Subjective tinnitus is the most common form of the disorder, but its neural origins remain rather elusive. Like any external acoustic percept, tinnitus is associated with neural activity in the central auditory system. Tinnitus-related activity may arise from abnormal cellular mechanisms in neurons of the central auditory system, or may result from aberrant input from the cochlea or non-auditory structures. Subjective tinnitus is often associated with peripheral hearing loss (Eggermont & Roberts, 2004; Nicolas-Puel et al., 2006) and many patients describe tinnitus as a sound in one or both ears. Therefore, it has been thought for many years that the tinnitus-related neural activity must also originate from the cochlea. Clinical observations however indicate the importance of central mechanisms. In patients that underwent sectioning of the eighth cranial nerve as part of retro-cochlear tumor surgery, tinnitus actually arose in 34% of the cases (Berliner et al., 1992). Surgically severing the connections between the cochlea and the brain has also been applied to provide relief of the tinnitus. This procedure was however not successful in 38-85% of cases (Kaltenbach et al., 2005).
Evidence for changes in the firing patterns of neurons in the central auditory system (Figure 3) as a possible substrate of tinnitus is mainly based on animal models. Noise trauma and ototoxic drugs, which are known to cause peripheral hearing loss and tinnitus in humans, result in behavioral responses in animals that seem consistent with the presence of tinnitus. One of these substrates is a change in the spontaneous firing rates of auditory neurons (Eggermont & Roberts, 2004). Since an increased (stimulus-driven) firing rate in auditory neurons typically corresponds to the presence of sound source, an increased spontaneous firing rate could also lead to an auditory percept, i.e., tinnitus. Alternatively, the temporal pattern of spontaneous neural activity could change by increased synchrony of activity across auditory neurons (Noreña & Eggermont 2003; Seki & Eggermont, 2003). The third candidate is a reorganization of the tonotopic map in auditory neurons in the central auditory system (Eggermont 2006; Mühlnickel et al., 1998; Seki & Eggermont, 2003). Although such changes themselves may not directly correspond to tinnitus, they may contribute to abnormal neural activity. For example, cortical reorganization may lead to the over-representation of frequencies corresponding to the frequency at the edge of a high-frequency sloping hearing loss (Eggermont, 2006; Rajan & Irvine, 1998). In other words, auditory-nerve or lower-brainstem neurons that are tuned to an edge frequency could be excessively projected to a region of the auditory cortex. fMRI can be used to identify mechanisms that underlie the generation of tinnitus in humans. Section 4.3 describes the application of fMRI and discusses the outcomes of selected studies.
Although there is much evidence for neural changes based on animal models, there is little corresponding evidence for tinnitus in humans. fMRI is expected to be sensitive to changes in the overall neuronal activity and may reveal changes in neural activity patterns. Yet, given the temporal resolution of typical fMRI studies, changes in synchronous neural activity or spontaneous firing rates may not be apparent from fMRI data. The BOLD signal is a relative measure of tissue properties and so ‘activity’ cannot be measured directly nor quantified. Instead, fMRI paradigms rely on modulating the hemodynamic response, reflected by neural activity in two (or more) conditions, which allows them to detect differences between states (Ogawa et al., 1990).
Given that tinnitus tends to be a continuous phantom percept, several paradigms have been applied to modulate it in some way. One approach examines sound-evoked responses in tinnitus and control participants, while the second method relies on the ability of a subgroup of people with tinnitus to manipulate their experience by somatic modulation. Examples of include forceful head and neck contractions (Levine, 1999), oral facial movements like jaw clenching or jaw protrusion (Pinchoff et al., 1998), electrical stimulation of the median nerve (Møller et al., 1992) and cutaneous stimulation (Cacace et al., 1999a; 1999b). Other paradigms that could be considered are i) the rapid change of gaze or tonic lateral gaze causing or modulating tinnitus, ii) pharmaceutical interventions that causes a temporal change of the tinnitus (e.g., lidocaine), and, iii) residual inhibition, which refers to the reduction of the loudness of tinnitus following the offset of an external acoustical stimulus (Roberts, 2007). A review of these paradigms is given by Lanting et al. (2009). Here, we highlight the contribution of fMRI to the current understanding of central mechanisms of tinnitus and explain how interpretations of neuroimaging data may help developing a better understanding of its diverse pathophysiology and of the functional interplay between sensory, cognitive and affective systems.
Compared to non-tinnitus controls, people with unilateral tinnitus initially appeared to show an abnormally small signal change in the inferior colliculus contralateral to the tinnitus percept, but not ipsilateral (Melcher et al., 2000). These findings were interpreted as evidence that tinnitus corresponds with abnormally elevated baseline neural activity on the contralateral side. When an external stimulus is presented, the hemodynamic response reaches saturation, resulting in a reduced difference between the ‘sound’ and ‘silence’ conditions. A potential confound to these observed patterns of sound-related activity was the influence of ambient background noise. Hence, in a new experiment, this variable was examined by repeating the study with the noisy helium pump either switched on or off (Melcher et al., 2009). All participants had normal hearing. With the pump off, the tinnitus group showed greater stimulus-evoked activation of the inferior colliculus than the non-tinnitus group. With the pump on, the tinnitus group showed reduced activation compared to controls. This result indicates some degree of auditory response saturation in the brains of people with tinnitus. A possible explanation is that a compensatory mechanism of central auditory gain has increased the response amplitudes close to their maximum limit. Lanting et al. (2008) used a sparse temporal sampling paradigm, minimizing the interaction between the scanner noise, the noise of the pump and auditory brain activity. Stimuli consisted of monaural dynamic rippled broadband noise stimuli at two intensity levels (40 dB and 70 dB SPL). Responses were measured at the level of auditory cortex and inferior colliculus of subjects with unilateral tinnitus and near-normal hearing. These were compared with those of people without tinnitus. This study demonstrated abnormally elevated responses in the inferior colliculus in people with tinnitus subjects compared to controls. Figure 4 shows increased sound-evoked responses, a reduced response lateralization (i.e., stimuli presented to the contralateral and ipsilateral ear gave roughly the same signal change) and a disturbed intensity level dependency in subjects with tinnitus compared to subjects without tinnitus at the level of the inferior colliculus (Lanting et al., 2008).
In summary, the first study by Melcher (Melcher et al., 2000) found decreased responses in inferior colliculus on the opposite side to the unilateral tinnitus percept, while the other two studies found overall increased responses (Lanting et al., 2008; Melcher et al., 2009). At first sight, these three results appear to be contradictory, but a key methodological difference may account for the discrepancy. While Lanting applied sparse temporal sampling, Melcher scanned continuously. Therefore, Melcher’s experiments were performed in high levels of ambient scanner noise which may have caused inferior colliculus to respond excessively to this source of background noise and hence reduce the BOLD response to the experimental sound stimuli. The three papers are thus consistent with the interpretation that the inferior colliculus of people with tinnitus displays a disproportionate response to sound, whether it is ambient sound or experimentally controlled sound.
Although sound-evoked responses appear elevated in people with tinnitus, hyperacusis could be an important alternative explanation. This condition is defined as an abnormal sensitivity to sound and is manifest as a diminished level of sound tolerance possibly elevating sound-evoked responses in the auditory brain. Hyperacusis is often co-morbid with tinnitus (Baguley, 2003). A recent study assessed the neural correlates of hyperacusis, recruiting 13 people with tinnitus and 14 people with no tinnitus (Gu et al., 2010). Hyperacusis was assessed using minimum loudness discomfort levels and a validated questionnaire. Severity of sound level tolerance (i.e. hyperacusis) was correlated with elevated activity in the auditory midbrain, thalamus, and primary auditory cortex, despite a fixed sound stimulation level. Sound-evoked activity in the inferior colliculus and medial geniculate body was more associated with hyperacusis than tinnitus, while in primary auditory cortex it was more associated with tinnitus than hyperacusis. Previous findings of increased sound-evoked responses in the inferior colliculus of people with tinnitus (Lanting et al., 2008; Melcher et al., 2009) are most likely to be due to the hypersensitivity to an external sound stimulus than tinnitus per se.
In conclusion, although fMRI has shown general increases in sound-evoked activity in tinnitus, it is more likely to be associated with hyperacusis. Any hypothesized reorganization of tonotopic maps in the auditory cortex has not yet been demonstrated, mostly due to the fact that tonotopic mapping is a relative new field. With the increasing availability of ultra high-field fMRI, this is a promising area for further research.
In addition to auditory nerve input, dorsal cochlear nucleus is also innervated by fibers from various parts of the somatosensory system, so this structure is a multi-modal processing station that is considered part of the extralemniscal auditory pathway (Figure 3). Parts of inferior colliculus, medial geniculate body and non-primary auditory cortex are all characterized by sensitivity to somatosensory stimuli (Eggermont & Roberts, 2004). As mentioned in Section 4.3, somatosensory manoeuvers can elicit or modulate tinnitus. Somatic modulation of tinnitus may therefore be considered a special case of multisensory integration—a phenomenon, in which one (sensory) modality influences another. The observation that some people with tinnitus can modulate the loudness or pitch of their tinnitus might be explained by changes in normal multisensory integration. Noise-induced hearing-loss, for example, has been reported to alter the normal somatosensory input (Shore et al., 2008). Somatosensory input to the cochlear nucleus increased after hearing loss, changing the balance of somatosensory and auditory input at the level of the brainstem, which in turn might be the neural correlate of somatosensory modulation of tinnitus.
To investigate this proposed model of cross-modal integration, Lanting et al. (2010) studied two groups: people with tinnitus and matched controls. Tinnitus participants were selected according to their ability to change the psychoacoustical characteristics of their tinnitus by jaw protrusion. Unimodal fMRI responses to sound stimuli and jaw protrusion alone were measured, as well as the response obtained in multimodal condition. To minimize motion artifacts, a sparse sampling method (inter-scan interval = 10 s) was used in which the protrusion started 4 s after the beginning of the acquisition and ended 1.5 s before the next acquisition. Since the jaw was always fully relaxed during scan acquisitions, the task did not degrade data quality. Results showed responses to jaw protrusion throughout the central auditory pathway, in both groups. Auditory responses to jaw protrusion presumably account for the modulation of tinnitus by these manoeuvers. Interestingly, responses in cochlear nucleus and inferior colliculus were greater in people with tinnitus than in controls, suggesting an abnormal auditory-somatic interaction in the patient group. This increased sensitivity to somatosensory inputs may correspond to neuroplastic changes, triggered by pathological change of either the somatosensory or auditory input pathways to dorsal cochlear nucleus (see Figure 3 and also Shore, 2011) that, in turn, lead to compensatory shifts in the balance of excitation and inhibition often attributed to tinnitus.
Hearing loss is the most common sensory deficit. Approximately 1.5% of children and 60% of over-70s have hearing loss and, as the average age of the population increases, the social, economic and psychological impact of hearing loss is likely to become more of a burden. Hearing after mild to moderate deafness can be regained through acoustic amplification with a hearing aid, whilst more severe to profound hearing loss can be partially restored by replacing the non-functioning inner ear with a CI.
A CI is a surgically-implanted auditory prosthesis that provides a sense of hearing to deaf individuals. As part of a preoperative clinical evaluation, computerized-tomography (CT) and MRI are both routinely used to assess CI candidates and to aid pre-operative planning and patient counselling. CT provides excellent structural information about bony detail and is useful for assessing congenital malformations that would affect the course of the facial nerve. MRI on the other hand can visualize soft-tissue with high resolution and is used to evaluate the integrity of the auditory nerve and the patency of the cochlea. Traditional CI candidates were adults with such profound hearing loss they could not benefit from any acoustic hearing. Over the past 25 years CI candidature has been relaxed, to include children and infants in whom auditory deprivation has the most severe developmental consequences. More recently, CIs have been used to treat individuals with some residual acoustic hearing in one or both ears. To date, nearly 200,000 individuals worldwide have been implanted with a CI, approximately half of those being children.
Although many CI users are able to develop age-appropriate communication and language skills without the need for visual cues, outcomes vary considerably between individuals. A number of factors have been found to correlate with speech outcomes following cochlear implantation (Blamey et al., 1996; Miyamoto et al., 1994). Pre-linguistically deaf adults generally perform more poorly than post-linguistically deaf adults (Busby et al., 1993a; 1993b; Eddington et al., 1978), whilst a long duration of deafness prior to implantation is generally detrimental to performance (O’Donoghue et al., 2000; Sarant et al., 2001). Also, it has been shown that auditory experience can improve clinical performance of cochlear implant users (Blamey et al., 1996; Gantz et al., 1993; Rubinstein et al., 1999). However, individuals with remarkably similar hearing experiences can often have markedly different speech comprehension skills. For example, older pre-linguistically deaf children can occasionally succeed very well in understanding speech, whilst the majority do not (Schramm et al., 2002). Therefore, whilst a number of factors that influence the clinical performance of cochlear implants have been identified, the field currently lacks accurate predictors of cochlear implant performance that can be applied to an individual. Such a prognostic tool, could help direct limited health-care resources towards effective clinical care.
Current national guidelines within the UK suggest that unilateral cochlear implantation is recommended for people with severe to profound hearing loss who do not receive adequate benefit from an acoustic hearing aid. Within this group, simultaneous bilateral cochlear implantation is recommended for i) pre-linguistically deaf children, ii) individuals who are blind, and iii) those who are at risk of cochlear ossification. Therefore, sensitive prognostic tools could be used to mould this crude patient pathway to meet the needs of individual patients, whilst providing realistic indicators of potential outcomes and requirements for rehabilitation.
Behavioural assessments of hearing with a CI are difficult to assess in very young children, as well as individuals with significant physical and/or developmental delay. Therefore, some investigators have focused on developing objective, non-invasive measures of CI function. Electrically-evoked compound action potentials (ECAPs) and electrically-evoked auditory brainstem responses (EABRs) are both routinely used to measure CI function at the level of the auditory nerve and brainstem, respectively. Whilst ECAPs and EABRs are useful clinical tools to ensure that the stimulus levels provided by the CI processor are within the range of perception, there is little evidence to suggest that they correlate well with speech and language outcomes (Miller et al., 2008).
Ultimately, CI users are likely to develop good speech and language skills provided accurate auditory information reaches the auditory cortex. For this reason, attempts have been made to assess functional activation of the auditory cortex with a CI. A number of methods can be used to assess cortical responsiveness to auditory stimulation. However, most of these methods have significant limitations when they are applied to a paediatric CI population.
Earlier in this chapter we discussed the advantages of fMRI in auditory research: a technique that permits the mapping of auditory responses with fine topographic detail, as well multiple observations over time, within both adult and paediatric populations. However, the surgically-implanted magnetic component of a CI device is incompatible with the high magnetic field generated by most MRI scanners. Traditionally, prior to considering MRI, it was necessary to surgically-remove the magnet after cochlear implantation. Recently it has been shown that CI patients can undergo MRI using a low (1.5) Tesla scanner, provided the implanted device is tightly bound before scanning (Crane et al., 2010). Patients in this study received an implant on one side. Although Crane and colleagues observed no magnet displacement whilst visualizing the contralateral- and part of the ipsilateral-internal auditory canal, the artefact surrounding the implanted device (~7x5 cm) precluded functional imaging of the cortex on that side. Similarly, magnetoencephalography (MEG) is limited by the magnetic field associated with current CI devices.
Electroencephalography (EEG) is also used to investigate auditory-evoked cortical potentials. However, the electrical artefact generated by the CI confounds these recordings. Although algorithms have been developed to minimize the effect of CI-generated electrical signals in EEG recordings (Debener et al., 2008), stimulus parameters are generally limited to short-duration sounds such as clicks (Gilley et al., 2006). Responses to longer sounds, including speech and language stimuli, are difficult to extract from the electrical artefact. Functional activation of the auditory cortex can also be assessed using PET that involves the use of ionizing radiation. Due to the cumulative effects of radiation exposure, this technique is not ideal for longitudinal studies, particularly in children.
Sevy et al. (2010) recently demonstrated a relatively novel imaging technique called ‘near-infrared spectroscopy’ (NIRS) could be used to assess cortical responses to auditory stimulation in children with cochlear implants. NIRS measures relative changes in the tissue contraction, oxy-haemoglobin, deoxy-haemoglobin and total haemoglobin in the cerebral cortex through beaming infra-red light. One of the greatest advantages of this approach is that the optical signals from the auditory cortex are not influenced by the magnetic fields that plague fMRI and MEG recordings. Unlike EEG, NIRS is not affected by CI-generated electrical artifacts. Subjects experience no discomfort with NIRS, as it is a non-invasive test that only requires contacting small optical fibre tips onto the scalp. There are no known risks associated with NIRS. Therefore, unlike PET, it is suitable for multiple measurements in paediatric populations. Although, NIRS provides sufficient resolution to measure evoked responses that are confined to auditory regions of the brain, this spatial resolution (within 1-2 cm of the activated area) is less than that of fMRI. Regardless, compared with alternative techniques, NIRS seems highly suited to assess the responsiveness of the auditory cortex before and after cochlear implantation in paediatric populations.
Animal neurophysiology has demonstrated that when sensory neurons lose their input (e.g. vision, touch or hearing) they do not become ‘silent’. Instead, they change their responsiveness. In the case of (noise-induced) sensorineural high-frequency hearing loss, neurons previously responding to high frequencies respond to low frequencies (Rajan et al., 1993; Robertson & Irvin, 1989). In the case of complete deafness, neurons become rewired to previously unconnected neurons from another sensory modality (Lomber et al., 2010; Rauschecker et al., 1995).
Cross-modal reorganization refers to the takeover of the processing of sensory information using new cortical pathways. It may be of clinical relevance for hearing-impaired individuals and individuals with a cochlear implant: some people with congenital hearing loss show ‘supra-normal’ visual abilities (Bavelier et al., 2006; Neville & Lawson, 1987). Furthermore, congenital deafness is known to induce rewiring of inputs within the auditory cortex (Allman et al., 2009; Auer et al., 2007; Finney et al., 2001; 2003; Levänen et al., 1998; Lomber et al., 2010; Nishimura et al., 1999) that plays a vital role in allowing us to process speech and naturally occurring sounds. Recently, Lomber et al. (2010) used cortical cooling techniques in deaf cats to suggest that enhanced visual performance is caused by cross-modal reorganization of the posterior auditory cortex.
Currently, it is not clear whether this rewiring is beneficial to hearing with a cochlear implant (potentially acting as an instructive map to ‘train’ the reinstated hearing pathways) or detrimental to hearing (by taking away resources from the auditory brain). Regardless, it remains possible that cross-modal reorganization following deafness may prove to be a clinically-useful predictor of performance following cochlear implantation. Lee et al. (2001) used F-18 fluorodeoxyglucose PET to examine resting-state brain activity in pre-lingually deaf individuals. Results indicated an association between hypometabolism in the auditory cortex and the amount of improvement in hearing after cochlear implantation. They speculated that the hypometabolism may be related to cross-modal plasticity and that the individual status of auditory rewiring might be used as an effective predictor of an individual’s speech performance following cochlear implantation. More recently, Rouger et al. (2011) demonstrated cross-modal responses in post-lingually deaf adults with cochlear implants engaged in a visual speech reading task using PET brain imaging. Furthermore, they suggested that cross-modal reorganization diminishes in post-lingually deaf adults with CIs with increasing time after implantation (up to at least one year later), towards the levels seen in normal-hearing individuals (Rouger et al., 2011). Doucet et al. (2006) compared cortical evoked potentials between adults with good and poor speech outcomes following cochlear implantation. They found that poor performers exhibited broader, anteriorly-distributed evoked potentials whereas the good performers showed significantly higher amplitudes of the P2 component over the primary visual cortex. Furthermore, Buckley et al. (2006) suggested that cortical responsiveness to visual and auditory evoked potentials accounts for a significant amount of the variability observed in speech perception performance in adults with cochlear implants with pre-lingual, but not post-lingual, hearing loss. Although, approximately half of cochlear implant recipients are children worldwide, currently, little is known about cross-modal brain reorganization during development.
While the results from a number of functional imaging studies in humans provide preliminary insights into plasticity within the central auditory pathway, more information is needed to understand the patterns of neural reorganization and the mechanisms underpinning those changes. This is an exciting time as novel imaging techniques, such as ultra-high field fMRI and NIRS, open new possibilities for investigating the effects of hearing loss, cochlear implantation and tinnitus on the developing and the mature brain. From our review, three key questions capture avenues of future research that are likely to impact on clinical practice: i) What are the separate effects of deafness, hyperacusis, tinnitus and cochlear implantation on central auditory plasticity, including cross-modal reorganization? ii) What are the perceptual consequences of that reorganization, and iii) How does brain reorganization affect and/or predict treatment outcome?
An ultimate goal for functional imaging is to provide the information necessary for evidence-based healthcare, such as targeting tinnitus treatments to specific neural sites of its generation and predicting outcomes following cochlear implantation. Currently, imaging of the central auditory pathway is not used for routine clinical purposes in Ear, Nose and Throat or Audiology clinics. It is our hope that research using imaging techniques may provide valuable insights into mechanisms underlying tinnitus and the variable speech outcomes following cochlear implantation in humans that may tailor our treatment towards the requirements of the individual.
1039total chapter downloads
Login to your personal dashboard for more detailed statistics on your publications.Access personal reporting
Edited by Dorina Papageorgiou
By Aymeric Guillot, Franck Di Rienzo and Christian Collet
Edited by Jill M. Norvilitis
By Dalia Mohamed Hassan and Hanan Azzam
We are IntechOpen, the world's leading publisher of Open Access books. Built by scientists, for scientists. Our readership spans scientists, professors, researchers, librarians, and students, as well as business professionals. We share our knowledge and peer-reveiwed research papers with libraries, scientific and engineering societies, and also work with corporate R&D departments and government entities.More about us