Auto-generated

This file is compiled by npm run compile-ba. Do not edit manually.

0.2 Abstract

Abstract Text


1.0 Introduction

1.1. The Auditory Dual-Stream Framework

The prefrontal cortex not only is the most important brain region for decision making, but also serves as the main hub for top-down attention. When we decide to look at the squirrel sitting on the table, it is our prefrontal cortex that first decides and then initiates the directing of the eyes and also the visual system to attend to the squirrel. While focusing on the squirrel, we hear the squirrel making noises - we then attend to the auditory stimuli as well. The question still remains which parts of the prefrontal cortex direct attention to the auditory stream. My thesis will examine the scope in which the prefrontal cortex directs attention top-down in different modalities. But first, we need to dive into the existing literature about the visual stream hypothesis.

1.1.1 The Visual Two-Streams Hypothesis

Lesion studies in macaque monkeys by Ungerleider and Mishkin (1982) set the start in developing the concept of parallel processing streams of the visual system in the brain. They described two distinct pathways that both originate in the primary visual cortex (V1): a ventral stream projecting toward inferotemporal cortex, processing objects such as a squirrel or the table. This stream is also called the ‘what’-stream and decodes object identity. The second, the dorsal stream, projects toward posterior parietal cortex and processes motion and object locations - also called the ‘where’-stream. Goodale & Milner (1992) later refined this framework, proposing that the dorsal stream is primarily involved in guiding how to interact with objects - while the ventral stream serves object perception and recognition. For our study it is relevant to ask whether the ‘how’/‘where’-stream not only performs visuomotor, but also audiomotor control over downstream modalities.

Critical for our study is recent work by Bedini and Baldauf (2021), identifying prefrontal hubs that perform top-down control over each stream. Based on evidence from functional and structural connectivity, they demonstrated a clear dissociation in functional connectivity: The Frontal Eye Field (FEF) - a core node of the Dorsal Attention Network (DAN)- shows predominant coupling with regions of the dorsal visual ‘where’-stream, while the anterior Inferior Frontal Junction (IFJa) - part of the Frontoparietal Network (FPN) - couples preferentially with the ventral visual ‘what’-stream. This connectivity-based division of labor was later supported by resting-state MEG data showing the same dissociation in oscillatory coupling and top-down directionality (Soyuhos, O., & Baldauf, D. (2022)). Together, these findings draw a clear picture of functional specialization in prefrontal top-down control.

1.1.2 From Wernicke’s Area to Auditory Dual-Stream Processing

In the auditory modality, the processing of input was thought to operate in a single cortical region, Wernicke’s area. It is located in the posterior left superior temporal gyrus (STG) and was considered the primary hub for auditory comprehension following Wernicke’s (1874) findings in aphasia - the missing ability to understand language. This model dominated neuroscience well into the twentieth century.
In the 1970s and 1980s, when neuropsychological evidence revealed that lesions to the left STG did not consistently produce comprehension deficits, but were instead more reliably associated with speech production deficits (Hickok & Poeppel 2007 - Nature). These findings changed the view, so that auditory processing could not be reduced to a single region and led to a fundamental re-evaluation of the cortical auditory organization.

The current view, developed most influentially by Hickok & Poeppel (2004, 2007) and Rauschecker & Scott (2009), puts the organization of the auditory cortex into two parallel processing streams analogous to those in the visual system. A posterodorsal stream projects from the superior temporal plane through parietal and premotor cortex and is associated with spatial processing and sensorimotor integration. An anteroventral stream projects from the STG forward through the temporal lobe toward inferior frontal regions and supports auditory object identification and semantic (speech) processing. Figure 1 illustrates this dual-stream architecture as proposed by Hickok & Poeppel (2007), mapping those two pathways onto a lateral view mainly of the left hemisphere. Section 3.2 ‘Selection of Regions of Interest’ will display in detail which regions most likely belong to which of both streams based on existing literature.

Figure 1. Schema of the auditory dual-stream model (adapted from Hickok & Poeppel, 2007). The dorsal pathway (blue) extends from the superior temporal plane toward posterior parietal and premotor cortex. The ventral pathway (purple) projects anteriorly through the temporal lobe toward inferior frontal regions. Both streams originate in the primary auditory cortex on the supratemporal plane.

1.2 The Gap, Top-Down Control of Auditory Streams

In environments with competing sounds, the brain cannot attend to all auditory input simultaneously. Top-down attention functions as a control mechanism that filters relevant information and suppresses distractors - therefore focussing on few inputs in a goal-directed manner (De Vries & Baldauf (2021) - Journal of Neuroscience). The key question is: Which prefrontal regions act as the top-down controllers of these auditory streams?

The division into dorsal and ventral pathways is well-documented(Ahveninen et al. (2006) - PNAS, Hickok & Poeppel 2007 - Nature), which already sets the path for following the question about the top-down attention mechanisms. Existing work has identified IFG subregions as frontal nodes in auditory processing, particularly BA44 and BA45 within the semantic pathway (Rolls et al. (2023) - Cerebral Cortex) and BA44 also along the dorsal pathway for affective prosody (Frühholz (2015) - NeuroImage). Hickok & Poeppel 2007 - Nature described a dorsal pathway including the Spt - a region at the parietotemporal boundary within the Sylvian fissure - as a sensorimotor interface, connecting anteriorly to Broca’s region and premotor cortex. Hickok & Poeppel 2007 - Nature. However, these findings remain specific to the language-related dorsal prosody processing or the semantic ‘what’-stream. Especially the auditory ‘where’-stream lacks a clear prefrontal controller. The question of whether the FEF and IFJa also direct the auditory domain in a similar fashion to the visual system remains unanswered.

1.3 Hypothesis, A supramodal organization

The preceding sections have outlined the missing piece: while the dual-stream architecture of the auditory cortex is well established, the prefrontal regions that coordinate these streams from the top-down remain under-explored. For the visual system, this question has been resolved - FEF and IFJa act as dissociable control hubs for the dorsal and ventral streams (Bedini & Baldauf (2021)); whether a comparable organization exists for the auditory domain is still unknown.

We hypothesize that the resting-state functional connectivity will show a clear dissociation of the auditory ‘what’ and ‘where’-streams: the auditory ‘where’-stream preferentially connects to the Frontal Eye Field (FEF), and the auditory ‘what’-stream to the anterior Inferior Frontal Junction (IFJa).

If confirmed, this dissociation would provide evidence for a supramodal organizational principle of the prefrontal cortex, with FEF and IFJa acting as domain-general hubs for top-down attentional control across sensory modalities. This study tests a first step toward that larger claim by examining whether FEF and IFJa connectivity profiles in the auditory domain mirror those established for vision.


2.0 Theoretical Background

The auditory dual-stream framework proposes that sound processing diverges into two functionally and anatomically distinct pathways after being processed by the primary auditory cortex.
The ventral ‘what’-stream decodes object identity and the dorsal ‘where’-stream processes spatial localization and sensorimotor integration. This organization mirrors the well-established visual dual-stream architecture. Romanski (2004) demonstrated in non-human primates that this division is not modality-specific, but ‘domain’-specific: the dorsolateral prefrontal cortex (dlPFC), including areas 46, 8a and 8b, receives converging spatial input from both posterior parietal (visual) and caudal superior temporal (auditory) cortices, while the ventrolateral PFC (vlPFC), including areas 45 and 12, receives object-related input from inferotemporal and anterior auditory association cortices (Figure 1, Romanski 2004, originally based on Arnsten (2003)). This domain-specific principle provides the theoretical foundation for the present thesis. The mapping of these macaque pathways onto human prefrontal architecture - specifically the FEF for the dorsal stream and the IFJa for the ventral stream - is elaborated in Sections 2.1. and 2.2, grounding on neuroimaging evidence (Bedini & Baldauf (2021)).

2.1 The Auditory What (“Ventral”) Stream

The auditory ‘what’-stream represents the more conceptually stable part of the dual-stream architecture, mostly tasked with transformation from auditory signals to semantic representations. The ‘what’-stream’s role in mapping sounds to meaning remains widely accepted as fundamental. This chapter discusses existing literature about the auditory ventral ‘what’-stream.

2.1.1 Grouping by Rolls (2022)

Throughout this thesis, I will refer to the groups proposed by Rolls et al (2022), that identified three large-scale networks using effective connectome analysis. These groups provide an anatomical reference for the regions discussed throughout the thesis. Their Group 1 consists of inferior STS regions (STSva, STSvp), anterior inferior temporal cortex (TE1a), temporal pole (TGd) and parietal PGi, aligning closely with the ventral stream described here. This group has effective connectivity (EC) directed toward Broca’s areas (44, 45, 47s), receiving input from high-level areas, has EC with memory-related regions, with the frontal pole and EC directed to parts of Group 2. This group includes a large frontal system with regions 44, 45, 47l, SFL and 55b, also including inferior temporal region TGv. This group has robust EC with superior (STSda, STSdp, STGa) and inferior regions as STSvp and the Peri-Sylvian Language Area (PSL). Finally, Group 3 is centered around superior temporal lobe and the temporoparietal junction with areas A5, STGa, STSda, STSdp, PSL, STV and TPOJ1. They show EC with some visual and auditory areas (Belt regions, A1, A4, A5), also with the frontal operculum and has strong EC directed towards 44 and 45.

2.1.2 Functional Definition

The primary objective is auditory object identification - the process of isolation of what is being heard (e.g. specific words, a dog’s bark or a piano) from a chaotic acoustic environment.
It operates as a sound-to-meaning starting with feature-extraction in the primary auditory cortex, which responds well to pure tones (Rauschecker & Afsahi (2023) - Journal of Comparative Neurology). Top-down attention in the ‘what’-pathway is mediated by the IFJ, mainly the anterior part. The IFJ is located at the junction of the inferior precentral sulcus (iPCS) and the inferior frontal sulcus (IFS), corresponding to BA6, BA8 and BA44 (Bedini & Baldauf (2021)). The IFJ is subdivided into an anterior and a posterior part:

  • anterior Inferior Frontal Junction (IFJa): Unlike the FEF, the IFJa processes non-spatial, object-based information and is part of the Frontoparietal Network (FPN) and functions as a mediator of top-down, feature-based attention, particularly in language-related networks (Ji et al., 2019, as cited in Bedini & Baldauf, 2021). It is also involved in working memory, and may encode information at a higher level of abstraction than the FEF. It can modulate DAN activity according to current task demands and toggle between DAN and VAN activity. Critically, IFJa receives EC from Groups 2 and 3 (Rolls et al., 2022), but not from Group 1 - suggesting that IFJa is preferentially accessed by frontal language output regions and superior temporal auditory areas, but not by the inferior STS network.
  • posterior Inferior Frontal Junction (IFJp - Control Seed): While the majority of neuroimaging literature treats the IFJ as a monolithic unit, recent structural and functional evidence suggests a clear dissociation between its anterior (IFJa) and posterior (IFJp) subdivisions (Bedini (2023) - Brain Structure). IFJp acts as a core node of the multiple-demand system, activated by generalized executive tasks (e.g., reasoning, math). We included IFJp specifically as a control region to validate the target specificity of the IFJa towards the auditory what-pathway.

2.1.3 Anatomical Trajectory

The anteroventral auditory ‘what’-stream originates in the Herschl’s Gyrus, extending through the Superior Temporal Gyrus (STGa), and along the Superior Temporal Sulcus (STS), Middle and Inferior Temporal Gyrus, to the anterior inferior frontal gyrus, the Broca’s areas 44, 45 and 47l (Ahveninen et al., 2006; Frühholz, 2015). Additionally, Soyuhos & Baldauf (2023) demonstrated that anterior IFJ has strong power coupling with the temporal cortex in delta and gamma frequencies.

2.1.4 Differences in Lateralization

The auditory ‘what-stream shows a left hemispheric dominance for language-related processing. Rolls (2022) - NeuroImage reports stronger left-hemispheric EC across what-stream regions STS, temporal pole, PGi and inferior frontal regions 44, 45 and 47l. The right hemisphere, however, still plays a complementary role in emotional and prosodic aspects of speech, activating right STG and IFG (Frühholz (2015) - NeuroImage), suggesting a division of labor in which the left hemisphere dominates semantic processing and the right contributes to the affective dimension of auditory objects.

2.2 The Auditory Where (“Dorsal”) Stream

In contrast to the ventral ‘what’-stream, the dorsal pathway has no single, universally accepted functional definition. Depending on the theoretical framework, the dorsal stream has been characterized as a spatial localization system (Rauschecker & Scott (2009) - Nature Neuroscience), a sensorimotor integration pathway for speech (Hickok & Poeppel 2007 - Nature), or a system for processing affective prosody (Frühholz (2015) - NeuroImage). These accounts are not mutually exclusive — they reflect different levels of description of the same anatomical pathway. For the purposes of this thesis, the dorsal stream is treated as a spatial-motor system subserving both spatial orienting and audiomotor integration, with the FEF as its prefrontal top-down control hub.

2.2.1 Spatial Processing and Audio-Motor Integration

The spatial-motor characterization of the dorsal stream focuses on converging evidence across species and methodologies. Early support came from non-human primate studies demonstrating posterior auditory cortex sensitivity to sound location (Rauschecker & Afsahi (2023) - Journal of Comparative Neurology), complemented by neuroimaging showing right-lateralized temporo-parietal activation for spatial processing and the clinical observation that spatial neglect is more severe after right-hemisphere damage.

Ahveninen et al. (2006) extended this account by reporting a double dissociation between anterior ‘what’ and posterior ‘where’ processing, with dorsal stream responses emerging at 70–150ms post-stimulus onset. Rauschecker (2011) further noted a cross-species shift: whereas non-human primates recruit the dorsal stream primarily for spatial localization, humans appear to repurpose it for sensorimotor integration in speech.

This sensorimotor dimension was elaborated by Hickok and Poeppel (2004, 2007), who proposed that the dorsal stream maps acoustic speech signals onto articulatory representations, projecting from the STG via the Sylvian parietal temporal area (Spt) to frontal regions including Broca’s area. Rather than contradicting the spatial account, this suggests that the dorsal pathway supports a broader spatial-motor function — localizing sounds in space and coupling that information with motor action plans.

2.2.2 Anatomical Definition and Evidence

The auditory dorsal “where”-stream originates posterior to Herschl’s Gyrus in the planum temporale and STG (Ahveninen et al. (2006) - PNAS), and extends through auditory association areas before reaching parietal and frontal regions. A4 and A5 have EC to motion-sensitive areas MT and MST, which connect to superior parietal regions (7AL, 7Am, 7PC), consistent with a dorsal “where”-stream (Rolls et al. (2023) - Cerebral Cortex).
Further, the stream is characterized by Rolls’ (2022) Group 3, with connectivity to auditory belt regions, motion-sensitive visual areas, and somatosensory cortex.

The Frontal Eye Field (FEF), located in the caudal middle frontal gyrus ventral to the junction of the superior precentral sulcus (sPCS) and superior frontal sulcus (SFS), serves as a primary prefrontal node of the dorsal stream. The FEF contains a full topographic map of contralateral space, predominantly encodes spatial information (cited from Bedini (2023) - Brain Structure: Wang et al., 2015) and is a core node of the Dorsal Attention Network (DAN), mediating covert spatial attention, oculomotor control, and spatial working memory (Bedini & Baldauf (2021)). It is structurally connected via SLF1 and SLF2 to posterior parietal (LIPd) and temporoparietal cortices, embedding FEF anatomically in the “where”-stream.

The inferior parietal cortex (specifically PF, PFop and PFcm) corresponds to the gyrus of the classical IPL (Baker (2018)) and forms a critical relay station within the “where”-stream. Consistent with both Rauschecker & Scott (2009) - Nature Neuroscience and Hickok & Poeppel 2007 - Nature/Hickok & Poeppel (2004) - Cognition, the “where”-pathway projects from posterior superior temporal cortex (pST) through parietal cortex before reaching prefrontal regions. These inferior parietal areas function as higher-level coordinator between auditory and somatosensory areas such as OP1-4 and FOP1 (Glasser et al. (2016) - Nature). Additionally, posterior parietal areas may serve as a relay station to the premotor cortex (PMC).
Importantly, the IPL does not appear to be driven by acoustic stimuli (Rauschecker & Scott (2009) - Nature Neuroscience), rather the angular gyrus is involved in higher-order speech processing - along with clear prefrontal activation. This suggests, the inferior parietal cortex is involved in domain-general, language-related speech comprehension, rather than low-level auditory processing (Rauschecker & Scott (2009) - Nature Neuroscience).
Superior parietal regions 7AM, 7PC, 7AL receive direct auditory input from A4 and PBelt and may support tracking moving objects in space (Rolls et al. (2023) - Cerebral Cortex).
Finally, A5 shows additional connectivity to IFJa and IFSp (Rolls et al. (2023) - Cerebral Cortex), suggesting a partial overlap between “where” and “what” - a boundary that will be explored in the analyses in Chapter 4.

2.2.3 The “Where”-Stream as a Spatial-Motor System

Taken together the evidence reviewed in 2.2.1 and 2.2.2 suggests that the auditory “where”-stream does not map on one functional label only, but rather has a spatial-processing (Rauschecker & Scott (2009) - Nature Neuroscience) and a sensori-motor-integration (Hickok & Poeppel 2007 - Nature) function and reflect different levels within the same dorsal pathway.
Spatial processing in the temporo-parietal cortex is predominantly right-lateralized, consistent with the well-documented asymmetry of spatial neglect in patients with right-hemispheric damage (Rauschecker & Scott (2009) - Nature Neuroscience). In contrast, Hickok & Poeppel’s dorsal language model that links area Spt to Broca’s area is strongly left-dominant (Hickok & Poeppel 2007 - Nature). This hemispheric dissociation suggests that the “where”-stream operates in parallel in both hemispheres with different functional emphases: right-lateralized spatial and motion processing, and left-lateralized sensorimotor integration.
Further findings support this account, with caudal auditory belt area receiving somatosensory input (Rauschecker & Scott (2009) - Nature Neuroscience) and the inferior parietal cortex - especially PF, PFcm, PFop - serving as an interface between auditory, somatosensory and motor signals.

A related question concerns the dorsal language pathway, Rolls et al. (2023) - Cerebral Cortex report that connections of PBelt, A4 and A5 with BA44 may form a language-related dorsal pathway. However, this is not in conflict with a spatial-motor property: The same stream seems to perform spatial processing and sensorimotor feedback in the right hemisphere and articulatory and phonological processing on the left.

For the purpose of this thesis, the “where”-stream is therefore treated as a spatial-motor system: primarily organized around spatial and motor processing, with a left-dominant component that extends into speech production. The FEF serves as a primary prefrontal node, embedded anatomically in the DAN and structurally connected to the posterior parietal regions.


3.0 Methods

3.1 Data Acquisition & Preprocessing

3.1.1 Dataset and Participants

This study utilized resting-state functional magnetic resonance imaging (rs-fMRI) data from the S1200 release of the Human Connectome Project (HCP). To ensure high data quality and completeness, a strictly defined sub-group of healthy young adult subjects was selected. This specific cohort was chosen because these subjects completed all four rs-fMRI runs, and their data were reconstructed using the latest r227 reconstruction algorithm, which corrects for specific phase-encoding artifacts present in earlier releases.

3.1.2 Data Acquisition

Imaging data were acquired using a customized Siemens 3T Connectome Skyra scanner. The rs-fMRI data were collected utilizing a gradient-echo EPI sequence with the following standard HCP parameters: a repetition time (TR) of , an echo time (TE) of , a flip angle of , and an isotropic spatial resolution of . Each subject completed two sessions on separate days. Each session consisted of two 15-minute runs with opposing phase-encoding directions (Right-to-Left and Left-to-Right), resulting in 1200 frames per run and a total of 4800 time points per subject. During the acquisition, subjects were instructed to keep their eyes open and maintain fixation on a red crosshair on a dark background.

3.1.3 Preprocessing

The raw fMRI data underwent preprocessing via the standardized HCP Minimal Preprocessing Pipelines. This first pipeline performs spatial distortion correction, motion correction, and registration to the structural final brain mask by reducing the bias field and normalization. Subsequently, the data were transformed from native mesh to fs_LR registered 32k mesh (2mm average vertex spacing).

To account for noise and motion artifacts, the data were denoised using ICA-FIX (Independent Component Analysis-based X-noiseifier). This automated method decomposes the data into independent components and regresses out those identified as structured noise (e.g., cardiac, respiratory, and head movement artifacts) without applying aggressive global signal regression.

3.2 Selection of Regions of Interest (ROIs)

To test the hypothesis of supramodal prefrontal control over auditory processing, a comprehensive set of 36 target ROIs and 2 primary seed regions (plus one control seed) per hemisphere was defined. All regions were identified utilizing the multimodal cortical parcellation (HCP-MMP1.0) by Glasser et al. (2016) - Nature. This atlas provides superior neuroanatomical precision by integrating structural, functional, and connectivity data.

The selected ROIs are categorized into functional networks reflecting the auditory dual-stream architecture and their respective prefrontal controllers. The large sample size (N=812) provides sufficient statistical power to detect moderate connectivity effects across all ROI pairs after multiple comparison correction

3.2.1 The HCP-MMP1 Atlas

Before defining specific regions, the choice of the underlying atlas must be justified. Unlike traditional parcellations based solely on cytoarchitecture (e.g., Brodmann areas), the HCP-MMP1.0 atlas integrates four distinct modalities:

  1. Cortical Myelin Content: Identified via T1w/T2w ratios.
  2. Cortical Thickness: Measuring structural differences.
  3. Task-fMRI Activation: Pinpointing functional hubs.
  4. Resting-State Functional Connectivity: Mapping intrinsic networks.

This multimodal approach is essential for this study because it allows for the differentiation of functionally distinct areas that appear homogenous in classical maps. For instance, it enables the isolation of the IFJa from the surrounding prefrontal cortex, which is important for the auditory dual-stream architecture.

3.2.2 Prefrontal Seed Regions (The Conductors)

We selected distinct prefrontal control hubs based on the functional dissociation previously established in the visual system by Bedini & Baldauf (2021). This selection tests whether these visual control architectures map onto auditory processing.

  • Frontal Eye Field (FEF): A core node of the Dorsal Attention Network (DAN), hypothesized to exert top-down control over the spatial auditory “Where”-stream.
  • Anterior Inferior Frontal Junction (IFJa): Part of the Frontoparietal Network (FPN), hypothesized to mediate feature-based attention and semantic processing within the auditory “What”-stream Bedini & Baldauf (2021).
  • Posterior Inferior Frontal Junction (IFJp) - Control Seed: Included strictly as a methodological baseline. IFJp is associated with the Multiple-Demand (MD) system for general-purpose executive tasks Bedini & Baldauf (2021). Factoring out its variance allows for a double dissociation, ensuring that the observed control over the auditory network is highly specific to the IFJa.

3.2.3 The Dorsal Pathways (Where & How)

The dorsal pathways process spatial localization (Where) and sensorimotor integration (How) (Hickok & Poeppel 2007 - Nature). ROI selection relies on the connectivity profiles outlined by Rolls et al. (2023) - Cerebral Cortex and recent functional data by Dureux (2024).

3.2.3.1 The Spatial-Orienting Network (Where-Stream)

This network tracks moving auditory objects and directs spatial focus.

  • 7AL, 7Am, 7PC: Superior parietal areas receiving direct input from early auditory regions, crucial for auditory spatial attention and motion tracking Rolls et al. (2023) - Cerebral Cortex.
  • PF, PFop: These areas correspond to the Inferior Perietal Lobule (Baker (2018)) and according to Rauschecker & Scott (2009) - Nature Neuroscience the IPL plays an important role in the dorsal auditory pathway (Glasser et al. (2016) - Nature)
  • PFcm: Defined by Glasser et al. (2016) - Nature as a heavily myelinated inferior parietal region with strong somatosensory properties along with OP1-4 and FOP1. Although recent evidence indicates a lack of direct auditory responsiveness (Dureux et al., 2024), it was strategically included due to its anatomical position bridging the auditory-motor interface (OP4) and multimodal integration hubs (PSL).
  • MT, MST: Classical supramodal motion hubs. They receive effective connectivity from auditory areas (A4/A5) to enable the tracking of auditory movement in space Rolls et al. (2023) - Cerebral Cortex.

3.2.3.2 The Motor Interface (How-Stream)

This interface translates auditory representations into articulatory motor plans (Sound-to-Motor mapping, Hickok & Poeppel (2004) - Cognition, Hickok & Poeppel 2007 - Nature)

  • 55b: A premotor hub exhibiting responses to vocal stimuli being assigned to a language area. Interesting because 55b is a neighbor of the prefrontal seed region FEF and Glasser et al. (2016) - Nature; Dureux (2024).
  • 44 (Broca’s pars opercularis): According to Rolls et al. (2023) - Cerebral Cortex Area 44 might be forming a dorsal auditory pathway with PBelt, A4, A5 and motion areas MT and MST.
  • OP4, FOP1, FOP2, FOP3, 43: According to Frühholz the frontal operculum is connected to posterior STG and anterior IFG via dorsal pathways Frühholz (2015) - NeuroImage. OP4 specifically demonstrates exclusive sensitivity to human vocalizations Dureux (2024). Area 43 was included due to its anatomical position between the FOP and premotor cortex, potentially serving as an interface along the auditory-motor-pathway.
  • SCEF (Supplementary and Cingulate Eye Field): A medial frontal cingulate area anatomically adjacent to the FEF. Despite responding almost exclusively to vocalizations (Dureux (2024)), its medial frontal anatomy and oculomotor system affiliation place it within the dorsal network. Dureux (2024) further groups SCEF functionally with premotor area 55b, OP4, and Broca’s areas 44 and 45 in the same vocalization-selective cluster.

3.2.4 The Ventral Pathways (What)

The ventral stream decodes auditory object identity, semantics, and speech perception.

3.2.3.1 Ventral Semantic and Identity Network

  • STGa, STSda, STSdp, STSva, STSvp, TA2: The semantic core of the ventral pathway processing complex auditory objects. The STS complex integrates vocal inputs with facial motor representations to decode identity and message (Glasser et al. (2016) - Nature, Rolls (2022) - NeuroImage, Rolls et al. (2023) - Cerebral Cortex)
  • AVI: The Anterior Ventral Insula shows activations to auditory stimuli alongside inferior frontal regions 55b, IFJp, IFJa, IFSp, 44, 45, and OP4. AVI exhibits highly specific capabilities in distinguishing vocalizations from noise, extending the semantic network into insular evaluation regions Dureux (2024).
  • 45, 47l, IFSp: The frontal termini of the ventral meaning pathway. Area 45 (Broca’s pars triangularis) and 47l handle semantic selection, while IFSp shows responses to vocalization along with 44, 45, IFJp, OP4 and 55b ; Rolls et al. (2023) - Cerebral Cortex, Dureux (2024)

3.2.4.2 Anterior Temporal Semantic Regions

  • TE1a, TGd, TGv (Temporal Pole): Following the effective connectivity analysis of Rolls (2022) - NeuroImage, these areas represent the semantic integration hubs of the temporal lobe. We intentionally differentiated between these regions to test the specificity of prefrontal control:
    • TE1a & TGd (Group 1): Associated with an inferior, visual-semantic system. Their lack of effective connectivity with the IFJa makes them ideal candidates to evaluate the boundaries of the IFJa-controlled auditory network.
    • TGv (Group 2): Integrated into a frontal system involving speech production and syntax. Its robust connectivity with the IFJa, FEF, and Area 55b makes it an interesting target for top-down modulation during linguistic and executive tasks.

3.2.5 Hierarchical Gateways and Connectors

These regions serve as routing hubs and convergence zones between early acoustic analysis and higher-order integration.

  • PBelt: A4, A5: The primary “routers” exiting the auditory core. PBelt and A4 show effective connectivity to parietal regions 7AL, 8AM, 7PC. Rolls et al. (2023) - Cerebral Cortex suggests that PBelt, A4 and A5 might form a language-related dorsal pathway, adding them to the auditory where-stream. Glasser et al. (2016) - Nature parcellation on the other hand places A4 and A5 together with STSdp, STSda, STSvp, STSva, STGa, and TA2 in a region naming them auditory association cortex. This area rather belongs to the ventral what-stream, making a clear classification difficult. Ambiguities will be resolved and discussed in chapters 4.4 and 5. (wo auch immer ich das mache)   
  • PGi: Effective connectivity to semantic areas as STS, TGv, TGd and TE1a place PGi in the auditory what-pathway. Rolls (2022) - NeuroImage places PGi in Group 1 along with STSva, STSvp and TE1a and TGd.
  • PSL, STV, TPOJ1: Multimodal convergence zones (Group 3 networks) bridging auditory semantics with visual and somatosensory inputs (Rolls (2022) - NeuroImage, Rolls et al. (2023) - Cerebral Cortex) showing effective connectivity with PBelt, A4 and A5. While PSL is anatomically integrated into the semantic network, functional models suggest it acts as an abstract linguistic interface rather than a primary auditory responder Dureux (2024). That’s why we placed these regions in the semantic what-pathway for the following analyses. TPOJ1 shows weak responses to non-vocal sounds according to Dureux (2024).

3.2.6 Tables

Table 1: Prefrontal Seed Regions

KürzelVoller NameLocation (Stream)Quelle
FEFFrontal Eye FieldPrefrontal (Dorsal Attention)Bedini & Baldauf (2021); Salmi et al. (2009)
IFJaAnterior Inferior Frontal JunctionPrefrontal (Frontoparietal)Bedini & Baldauf (2021)
IFJpPosterior Inferior Frontal JunctionPrefrontal (Multiple-Demand)Bedini & Baldauf (2021)

Table 2: Where & How Stream (Dorsal)

KürzelVoller NameLocation (Stream)Quelle
7ALArea 7 Anterior LateralDorsal (Where - Spatial)Rolls et al. (2023)
7AmArea 7 Anterior MedialDorsal (Where - Spatial)Rolls et al. (2023)
7PCArea 7 Posterior CapsularDorsal (Where - Spatial)Rolls et al. (2023)
PFArea PF (Inferior Parietal)Dorsal (Where - Spatial)Baker (2018)
PFopArea PF OpercularDorsal (Where - Spatial)Rauschecker & Scott (2009)
PFcmArea PF Complex MedialDorsal (Where/Somatosensory)Glasser (2016)
MTMiddle Temporal AreaDorsal (Where - Motion)Rolls et al. (2023)
MSTMedial Superior Temporal AreaDorsal (Where - Motion)Rolls et al. (2023)
55bArea 55bDorsal (How - Motor Relay)Dureux (2024)
44Area 44 (Pars Opercularis)Dorsal (How - Motor)Rolls et al. (2023)
OP4Frontal Opercular Area 4Dorsal (How - Motor interface)Dureux (2024)
FOP1-3Frontal Operculum 1, 2, 3Dorsal (How - Motor planning)Frühholz (2015)
43Area 43Dorsal (How - Motor planning)Frühholz (2015)
SCEFSupp. & Cingulate Eye FieldDorsal (How - Cingulate Attn.)Dureux (2024)
A4Auditory Area 4Gateway (Sensory Router)Rolls et al. (2023)
PBeltParabelt ComplexGateway (Sensory Router)Rolls et al. (2023)

Table 3: What pathway (ventral)

KürzelVoller NameLocation (Stream)Quelle
A5Auditory Area 5Ventral (What - Semantic Gateway)Rolls et al. (2022)
STGaSuperior Temporal Gyrus Ant.Ventral (What - Identity)Glasser (2016)
STSda/dpSTS Dorsal Ant. / Post.Ventral (What - Semantic)Rolls et al. (2023)
STSva/vpSTS Ventral Ant. / Post.Ventral (What - Semantic)Glasser (2016)
TA2Area TA2Ventral (What - Semantic)Glasser (2016)
STVSuperior Temporal Visual AreaVentral (What - Multimodal)Rolls (2023)
TPOJ1Temp.-Par.-Occ. Junction 1Ventral (What - Convergence)Rolls (2023)
PGiArea PGi (Inferior Parietal)Ventral (What - Visual Semantic)Rolls (2022)
AVIAnterior Ventral InsulaVentral (What - Evaluation)Dureux (2024)
45Area 45 (Pars Triangularis)Ventral (What - Semantic)Rolls et al. (2023)
47lArea 47 LateralVentral (What - Semantic)Rolls et al. (2023)
IFSpInferior Frontal Sulcus Post.Ventral (What - Semantic)Dureux (2024)
TE1aArea TE1 AnteriorVentral (What - Periphery)Rolls (2022)
TGd / TGvTemporal Gyrus Dor. / Ven.Ventral (What - Periphery)Rolls (2022)
PSLPerisylvian Language AreaConnector (Linguistic Interface)Rolls (2023)

3.3 Matrix Construction

3.3.1 Functional Connectivity

Resting-state functional connectivity (RSFC) is defined as the temporal correlation between neurophysiological events in spatially distinct brain regions Friston (1994). In this study, RSFC is operationalized as the statistical dependency between BOLD (Blood Oxygen Level Dependent) signal time series, reflecting the intrinsic functional architecture of the brain during resting-state Biswal (1995). Unlike EC, which models directed causal influences, RSFC is a symmetric, undirected measure that captures the degree to which two regions show correlated activity over time, independent of any explicit task or stimulus.

3.3.2 Time Series Extraction

We utilized the preprocessed dense CIFTI timeseries from the HCP S1200 release. For each subject, the BOLD signal was spatially averaged across all vertices within each of the 360 parcels (180 per hemisphere) of the HCP MMP1 atlas (Glasser et al. (2016) - Nature). Mean BOLD time series were extracted for all seed regions (FEF, IFJa, IFJp) [WELCHE noch? 55b, 44, 45?], as well as all auditory target regions defined in section 3.2.

3.3.3 Connectivity Matrix Construction

Pairwise RSFC values were computed as Pearson correlation coefficients between the extracted BOLD time series of each seed region and all parcels of the HCP-MMP1 atlas - consistent with the HCP’s own connectivity analyses (Glasser et al. (2016) - Nature). This led to a connectivity vector for each subject representing the full RSFC profile for each seed across the cortex. All matrix construction steps were implemented using a custom MATLAB toolbox developed within the Baldauf Research Group. [TOOLBOX Ref - Daniel fragen]

3.3.4 Single Seed vs. Contrast Analysis

We used two complementary analysis modes. In the single-seed analysis, the functional connectivity of a given seed region (FEF or IFJa) was assessed against the mean whole-brain connectivity of that subject, detecting parcels that are significantly more correlated with the seed region than the average cortical region. In the contrast analysis, the connectivity profiles of two seed regions were directly compared, revealing parcels with preferential connectivity to one seed over the other.

3.3.5 Partial vs Full Correlation

When characterizing functional brain networks, a critical distinction needs to be made between full and partial correlation. Full Correlation calculates the pairwise statistical relationship between two regions A and B without controlling for other variables. Because any two regions that both correlate with a third region C will appear correlated with each other - therefore displaying indirect correlation without true connectivity. Partial Correlation on the other hand, estimates the relationship between A and B after regressing out the shared variance to all other simultaneously measured regions (e.g. region C, Marrelec 2006 - NeuroImage, Smith (2011)). As a result, the partial correlation connectome is considerably sparser than its full correlation counterpart, while retaining the most direct functional connections (Glasser et al. (2016) - Nature).

In this study, both measures are applied: full correlation provides a global view of each seed’s connectivity landscape, while partial correlation isolates direct connections after filtering out contributions from the other regions included in the analysis.

3.3.6 Statistical Testing: Wilcoxon Signed-Rank Test & FDR Correction

Statistical significance of RSFC estimates was assessed using the paired Wilcoxon signed-rank test, an alternative to the Gaussian-based paired t-test. This test was chosen because RSFC values derived from Pearson correlations do not necessarily follow a Gaussian distribution across subjects (Soyuhos, O., & Baldauf, D. (2022)) For the single-seed analyses, a one-tailed test was applied, evaluating whether connectivity with the seed region exceeded the mean whole-brain connectivity. For contrast analyses, a two-tailed test was used to determine whether connectivity was significantly higher for one seed than for the other.

Each seed region (FEF or IFJa) was tested independently; their connectivity profiles were not averaged before statistical testing, preserving each region’s functional fingerprint.

To control for the elevated risk of false positives arising from testing all 360 cortical parcels simultaneously, all p-values were corrected for multiple comparisons using the False Discovery Rate (FDR) procedure (Benjamini, Y., & Hochberg, Y. (1995)). In contrast to the more stringent Bonferroni correction, FDR correction maintains higher statistical power by controlling the expected proportion of false positives among all rejected null hypotheses, rather than the probability of any single false positives. A significance threshold of q < 0.05 (FDR-corrected) was applied in all analyses.

For visualization, significant p-values were converted to z-scores. The z-scores were summed across all seeds for which a given parcel reached significance, and then divided by the total number of seeds showing significance for that parcel. The normalization generates a score that weights connectivity strength by consistency of the effect across seeds.

3.4 Brain-Behavior Correlation

To validate our RSFC patterns performed in Section 4, we assessed whether individual differences in RSFC could predict performance on behavioral tasks that specifically test the auditory ‘what’ and ‘where’ pathways. This approach links the network architecture observed during rest to cognitive function during task engagement, contributing validity for the supramodal organization hypothesis.

3.4.1 Behavioral Subject Subsample

For the connectome-based predictive modeling, we used behavioral data of N = 371 participants from the HCP S1200 release. While the full resting-state cohort counts participants, restricting the behavioral analyses to the smaller subset guarantees complete data across all evaluated tasks. For each task, subjects were additionally excluded if they exceeded a head motion threshold of 0.15 mm framewise displacement or had missing behavioral scores, resulting in slightly varying effective sample sizes per analysis (see sections 3.4.2.1 and 3.4.2.2).

3.4.2 Behavioral fMRI Tasks

Three fMRI tasks were selected to evaluate the auditory dual-stream framework, defined in our study. The tasks were selected to enable a functional dissociation between semantic processing ‘what’, auditory-spatial decoding ‘where’, and low-level acoustic filtering, while acknowledging the limitations of the HCP task set.

3.4.2.1 Language Task (Story Comprehension)

In the story condition of the HCP Language Task, participants listened to short auditory stories adapted from Aesop’s fables, followed by a two-alternative forced-choice (2-AFC) question testing the semantic content of the stories (Binder (2011), Barch et al. (2013)). Two performance measures were extracted: Story Accuracy measures the proportion of correct responses, and since the vocabulary of the stories is simple, we observe a ceiling effect in which the sustained attention is measured rather than semantic access. However, Story Median Reaction Time captures the speed of semantic retrieval and therefore is used as the primary measurement for ventral ‘what’-stream efficiency. For this analysis, 20 subjects were excluded from the N = 371 subsample (18 due to head motion > 0.15 mm, 2 due to missing scores), yielding an effective N = 351. The Spearman correlation between behavioral scores and head motion was (), confirming that motion artifacts did not confound the results.

3.4.2.2 Working Memory Task (2-Back Place)

The HCP Working Memory Task consists of a visual N-Back design in which participants track sequential images and respond whether the current image matches the one presented two trials before (Barch et al. (2013)). The place subcondition, in which stimuli consisted of landscape and scene photographs, was extracted as a measure for visuo-spatial working memory (Barch et al. (2013)).
This task was chosen as the closest available test of spatial processing capacity, linked to the FEF’s dorsal stream. Because the HCP dataset does not include an auditory spatial task, this visual task tests the supramodal hypothesis. Participants might solve these tasks via object recognition rather than spatial navigation, which might reduce observed effects in the dorsal network. Additionally, since IFJa has been shown to be involved in working memory (Bedini & Baldauf (2021)), its connectivity may contribute to Place task predictions independently of any semantic strategy, making it difficult to cleanly attribute ventral model effects to a single mechanism. For this analysis, 19 subjects were excluded from the N = 371 subsample (18 due to head motion > 0.15 mm, 1 due to missing scores), yielding an effective N = 352. However, the Spearman correlation between behavioral scores and head motion was (), indicating a significant association between motion and Place Acc performance. This constitutes a potential confound and is discussed as a limitation in section 5.7.

3.4.2.3 Words-in-Noise (NIH Toolbox Noise Comparison)

The NIH Toolbox Words-in-Noise test assesses the ability to recognize spoken words in background noise by adaptively decreasing the signal-to-noise ratio across trials (Zecker (2013)). The task is a valid measure of low-level acoustic signal separation rather than high-level semantic processing. This test was included as a control paradigm to determine whether IFJa network connectivity generalizes across auditory tasks or whether acoustic signal separation relies on a functionally different network.

3.4.3 Connectivity-Based Behavioral Prediction

The connectivity values were derived from partial correlation matrices computed across the N = 371 subsample, reflecting direct functional coupling between ROIs with shared variance removed. The prediction model itself was fitted using linear correlation between the connectivity predictor matrix and the behavioural outcome variable. Separate models were computed for the ventral ROIs subset and the dorsal ROIs subset, enabling predictions specific to the streams.

We evaluated model performance using K-fold cross-validation (K = 371, leave-one-out, as the primary analysis; K = 10 for comparison), providing an R value (correlation) as the indicator of predictive accuracy. A significance threshold of p < 0.05 was applied for all tasks.


4.0 Results

Our analyses reveal a clear functional dissociation between the auditory where- and what-streams at the level of their prefrontal top-down regulators. Partial correlation of the two seed regions FEF and IFJa shows a distinct connectivity that aligns with the proposed supramodal organization. The FEF is selectively coupled to spatial-parietal and motor circuits, while IFJa is preferentially connected to the ventral semantic network and the Broca complex. This double dissociation is further supported by seed-specificity validations (FEF vs. 55b, IFJa vs. IFJp) and by the analysis of Broca’s areas 44 and 45.

All z-scores reported below are from partial correlation analyses unless otherwise stated. Full correlation results are discussed only where they show meaningful contrasts compared to the partial correlation pattern.

4.1 Global Connectivity Patterns

Before examining each stream individually, we assess the overall functional dissociation between the auditory ‘what’ and ‘where’ streams. Using full and partial correlation, we compare the connectivity profiles of the two prefrontal seed regions FEF and IFJa, and validate their anatomical specificity against neighboring control seeds.

4.1.1 FEF vs. IFJa: Validation of Prefrontal Seed Regions

To validate the anatomical specificity of our seed regions, we compared their partial correlation connectivity fingerprints across both hemispheres (Figure FEF vs IFJa full connectivity.png).

The functional connectivity analysis with full correlation reveals a clear double dissociation (Figure: FEF vs IFJa full corr.png) with FEF significantly coupling with spatial-parietal areas ((full corr.: OP4 , 7Am , 7PC , RH)) with exceptions of FEF coupling with TA2 (full corr.: , RH) stronger than IFJa and TA2 and 55b connecting stronger to IFJa than to FEF.

Then we applied partial correlation to partial out third party connectors to see a clear picture of the connectivity patterns within the auditory streams. The partial correlation analysis reveals that the most direct auditory connections to FEF are consistently rooted in the spatial orienting network. Both hemispheres show strong coupling with the inferior parietal cluster (PF, PFop, PFcm). However, the network shows a distinct right-hemispheric dominance in specific regions, consistent with widely accepted right-lateralization framework for spatial auditory processing (Hickok & Poeppel, 2007). While the left hemisphere dominates in FOP3 (), 7Am () and 7PC (), the right hemisphere demonstrates a clear dominance in PF ( vs. left ) and shows partial correlation with the perisylvian language area (PSL, ), suggesting a stronger specialized right-lateralized fronto-parietal integration for spatial auditory processing (Hickok & Poeppel 2007 - Nature).



Figure: Functional Connectivity: Left vs right hemisphere, full correlation

Data: 0 DATA for Bachelorarbeit



Figure: Functional Connectivity: left vs right hemisphere, part correlation
Data: 0 DATA for Bachelorarbeit

  • FEF: The FEF seed exhibits robust full correlation with the superior parietal lobe (7AL, 7Am, 7PC). This replicates the where-stream, similar to the visual stream Bedini & Baldauf (2021), together with auditory and motor regions such as A5, PBelt, FOP1, STV and PSL
  • IFJa: In contrast, the IFJa showed strong coupling with the anterior language network (HCP language task)

4.1.2 Topographical Validation: IFJp Control Seed

To validate the top-down control over the auditory what-stream is driven by the anterior subdivision of IFJ, a control partial correlation model substituted IFJa for IFJp, its immediate neighbor.
The control analysis confirms the functional dissociation of where and what stream. The FEF maintains its robust connectivity to the dorsal auditory and motor network with 23 significant regions in the left and 20 in the right hemisphere.

In the profound contrast substituting the what-hub with IFJp leads to a significant drop in the connectivity pattern of the semantic what-stream. The IFJp fails to communicate with the temporal lobe in comparison to the FEF.
This contrast provides the evidence that the prefrontal control hub over auditory identity processing is functionally exclusively anchored in the IFJa.

FEF vs. IFJp right hemisphere: part corr

FEF vs IFJp part left hemisphere:

4.2 Testing the “Where” Stream (FEF Connectivity)

To evaluate the precise anatomical pathways of the top-down auditory attention network, the functional connectivity was assessed using full and partial correlation, revealing a clear spatial-dorsal pattern that places the FEF within its established framework.

full correlation

partial correlation:

partial correlation brain:

4.2.1 Inferior Parietal Connection

In visual stream models, FEF couples strongly with the superior parietal lobule areas 7PC, 7AM and 7AL. However, partial correlation reveals a stark contrast, not showing significant coupling of FEF to superior parietal regions 7AL, 7AM and 7PC, although still coupling bilaterally with inferior parietal regions of PF (left , ; right , ), PFop and PFcm.

4.2.2 Premotor cluster (“How” Stream)

The strongest connections with FEF are the premotor and frontal opercular areas. The SCEF (left , ; right , ) and area 55b (left , ; right , ) form a bilateral motor-premotor cluster, along with FOP1 and area 43. Together these areas might form the “how” stream, proposed by Hickok & Poeppel 2007 - Nature, connecting spatial information with motor action plans.

4.2.3 Resolving Superior Parietal and Inferior Temporal Connections

A crucial topological misconception emerges when comparing full and partial correlation profiles. In the full correlation, the FEF exhibits strong coupling with superior parietal areas (e.g., 7Am: left ; 7PC: left ). However, in the partial correlation model, these connections vanish entirely. This demonstrates that FEF does not maintain full spatial connections with superior parietal lobule when partialling out auditory ROIs. For auditory-spatial information, FEF mainly recruits Inferior Parietal Lobule (PF/PFop) and motor-sensitive areas. Furthermore, temporal areas that appear significant in Z-scores in figure FEF vs IFJa part corr.png (TE1a, TA2, TGd) show no substantial effect size (), suggesting that the FEF remains functionally decoupled from the ventral “what” stream’s temporal regions.

4.2.4 Motion and Supramodal Integration

The FEF maintains direct functional access to spatial motion tracking. Partial correlation reveals bilateral coupling with the medial superior temporal area (MST: left , ; right , ). Additionally, multimodal convergence hubs STV (left , ; right , ) and TPOJ1 (left , ; right , ) show significant direct integration with FEF in both hemispheres (Rolls et al. (2023) - Cerebral Cortex).

The right hemisphere reveals an exclusive direct connection to the PSL (, ), which is entirely absent in the left hemisphere’s partial profile, indicating a right-lateralized pathway for abstract spatial-auditory processing (Rolls (2022) - NeuroImage).

4.3 Testing the “What” Stream (IFJa Connectivity)

To map the precise anatomical pathways of the top-down auditory identity network, the functional connectivity pattern of IFJa was evaluated using single-seed partial correlation. The approach filters out shared variance, showing direct functional coupling of the ventral what-stream to the IFJa.

4.3.1 Temporal-Semantic connections

The partial correlation analysis confirms that the IFJa robustly couples areas in the superior temporal sulcus (e.g. STSdp: left , ; right , and STSda: left , ; right , ). Additionally, the left hemisphere integrates STGa and STSva. This direct prefrontal-temporal pathway validates the IFJa as a top-down controller for decoding auditory object identity.

partial correlation:

4.3.2. Prefrontal Language Connections

Within the prefrontal cortex, the IFJa serves as a master hub for the semantic language network. IFJa bilaterally couples with the Broca’s areas 45 (left , ; right , ) and Area 44 (left , ; right , ), along with connections to 47l. Additionally, IFJa couples strongly with the dorsal language hub 55b (left , ; right , ) and FEF (left , ; right , ). This positions IFJa as a main connector and attention hub for the language network.

4.3.3. Gateways and Convergence

Partial correlation reveals bilateral coupling with auditory areas A4 (left ; right ) and A5 (left ; right ), along with right lateralized PBelt connections (). Additionally, robust multimodal convergence areas like STV (left ; right ) and TPOJ1 are deeply integrated into the IFJa network, functioning as a combination of auditory identity with other sensory modalities. Additionally, PSL couples bilaterally with IFJa in both hemispheres (, right: , left: 9.65)

4.3.4 The Spatial Dissociation

A fundamental topographical dissociation emerges when contrasting IFJa’s full and partial correlation patterns, especially in the parietal lobe. The results show a strong correlation pattern across the motor and spatial areas (55b, OP4, MT/MST), but dissecting the connectivity pattern with partial correlation reveals instead sparse, language-driven coupling. This confirms the IFJ’s role as a prefrontal hub for semantic, temporal pathways, segregating it from the spatial “where”-stream.

full correlation:

4.4 Resolving Ambiguities

During the analysis, some areas as A4 or PSL seem to have an ambiguous connectivity pattern. This chapter will dissolve ambiguities found in the literature and results of the RSFC analysis.

4.4.1 A4 and A5

Areas A4 and A5 exhibit connectivity profiles that position them as early gateways for the auditory pathways targeting robustly IFJa (A4: left z = 8.25, right z = 8.76; A5: left z = 11.63, right z = 6.96). Partial Correlation shows only in left hemisphere low coupling from A4 with FEF (, ). This connectivity pattern aligns with the Glasser parcellation

As noted in the ROI selection (Section 3.2.5), A4 and A5 were

  provisionally placed in the dorsal pathway pending empirical resolution.

   The present data provides that resolution…

According to Rolls et al. (2023) - Cerebral Cortex A4 and A5 could form a dorsal language-relates stream with PBelt and Broca’s area 44. A4 and A5 both exhibit effective connectivity to MT and MST (Rolls et al. (2023) - Cerebral Cortex), which connect further to parietal regions, therefore, according to Rolls, A4 and A5 might form a language related dorsal stream.
Glasser

A4

In Glasser et al. (2016) - Nature A4 is defined as functionally different from its neighbors:

  • in contrast to STV, A4 shows a weaker connectivity to visual area PCV
  • We identified auditory association cortex as a region mainly on the superior temporal gyrus and within the superior temporal sulcus that is activated in the LANGUAGE STORY, MATH, and STORY-MATH contrasts. It is strongly functionally connected with the inferior frontal gyrus, including areas 44, 45, and 47l. This auditory region likely becomes progressively less purely auditory and more multi-modal as one progresses inferiorly, anteriorly, and posteriorly (away from early auditory cortex, e.g. Main Text Figure 3). Indeed, functional connectivity with early auditory cortex progressively decreases along those directions. This region includes eight areas that we identify as A4, A5, STSdp, STSda, STSvp, STSva, STGa, and TA2.

A5
Rolls (2023) treats A5 as an ambiguous
Rolls 2022 puts A5 in group 3 with STGa, STS, da, STSdp, PSL, STV and TPOJ1. Those regions show connectivity to Brocas areas 44 and 45, to TGv and TGd
also EC to MT/MST

Connectivity

Projections

PFC targets

4.4.2 PSL the division

PSL presents a fundamental paradox within the auditory dual-stream framework. On one hand, Rolls et al. (2023) - Cerebral Cortex classify PSL as part of the auditory ventral what-stream, based on its effective connectivity along with TPOJ1, STV, TGv, TGd and PGi, as language-related semantic regions connecting to Broca’s area 45. On the other hand, Dureux (2024) demonstrates that PSL remained entirely unresponsive to all tested auditory stimuli - including vocallizations, non-vocal sounds and white noise.
In our analysis, PSL presents a complex, asymmetric connectivity profile. It is integrated into the left ventral “what” stream (L_IFJa: , , single seed), but simultaneously forms a strong right-lateralized connection with the right spatial “where” stream (R_FEF: , , single seed).
Our partial correlation data resolves this apparent contradiction. The massive prefrontal connectivity, combined with Rolls et al. (2023) - Cerebral Cortex’s language-related integration and acoustic silence (Dureux (2024)), suggests that PSL does not process low-level auditory input. Instead, it might function as a high-level convergence zone similar to TPOJ, which calls Rolls (2022) - NeuroImage a multimodal convergence region. PSL might link semantic identity (via left IFJa) with spatial attention (via right FEF) at an abstract, modality-independent level.

4.4.3 STV

[DRAFT — bitte prüfen und in eigene Stimme übersetzen]

The Superior Temporal Visual Area (STV) presents a nominative paradox: despite its name implying visual processing, it is consistently recruited in both auditory streams and classified by Rolls et al. (2023) - Cerebral Cortex as part of the language-related ventral network alongside PSL and TPOJ1 — with strong effective connectivity directed towards Broca’s areas 44 and 45.

Our partial correlation data reveals significant bilateral coupling of STV with both prefrontal seeds — assessed in separate single-seed analyses. In the IFJa single-seed analysis (section 4.3), STV couples directly with IFJa (left , ; right , ), after partialling out FEF and all other ROIs. Conversely, in the FEF single-seed analysis (section 4.2.4), STV couples directly with FEF (left , ; right , ), after partialling out IFJa and all other ROIs. In both cases, the coupling is direct — not mediated through the other seed — indicating that STV maintains independent functional connections with each controller.

This dual connectivity resolves the apparent paradox. STV does not process unimodal visual input in isolation; rather, it functions as a supramodal integration zone at the temporal–parietal boundary, binding auditory identity signals from the ventral stream with spatial and cross-modal information from the dorsal stream. This interpretation aligns with its anatomical position at the posterior superior temporal sulcus, a region consistently implicated in audiovisual and multimodal convergence (Rolls et al. (2023) - Cerebral Cortex, Rolls (2022) - NeuroImage). Its strong IFJa coupling suggests a primary affiliation with the “what” stream, while its weaker FEF connection positions it as a candidate gateway through which semantic identity representations gain access to spatial attentional modulation.

4.5 Functional Roles of Adjacent Areas

4.5.3 Specificity Validation: FEF vs. 55b

A parallel validation compared FEF with its immediate premotor neighbor, area 55b, using partial correlation to rule out nonspecific regional signals. Despite their direct anatomical neighbourhood, the two seed regions exhibit clearly dissociable connectivity patterns. In both hemispheres, the FEF maintains dominant coupling with parietal, motor areas and the ventral connectivity to TA2 and TE1a already shown in Fig FEF vs IFJa part corr.png. In contrast, area 55b dominates language-related connectivity to areas A5, PSL (left , right ), superior temporal lobe and Broca’s areas 44 and 45 (44: left , right ; 45: left , right ). suggesting it functions as a language-relay for the spatial-dominated prefrontal dorsal domain led by the FEF. This dissociation confirms that the spatial connectivity profile is specific to the FEF, and not general for the entire dorso-prefrontal region.


Notably, 55b maintains strong direct coupling with auditory association areas A5 (left: , right ) and A4 (right: ). These connections are mostly absent in the FEF profile (only in left hemisphere to A4, , ). This suggests that 55b serves as the primary auditory-language relay within the dorsal prefrontal domain, connecting acoustic input with the spatial attention network rather than the FEF itself.

Data: 0 DATA for Bachelorarbeit

4.5.4 Results of the Broca-Seed Validation

To evaluate recent anatomical categorizations proposing a functional subdivision of Broca’s area within the auditory dual-stream architecture (Rolls et al. (2023) - Cerebral Cortex), the connectivity patterns of Areas 44 and 45 were independently assessed.

Applying partial correlation to isolate direct functional pathways reveals highly divergent connectivity profiles, highlighting the distinct specializations within the Broca’s area.

Area 45: The Ventral Anchor
Area 45 exhibits a strong integration into the ventral what-stream. The connectivity profile extends massively into the temporal lobe (e.g., STSdp: left , ; right , ) and demonstrates strong coupling with the PSL (left , ; right , ). This robust temporal-prefrontal coupling firmly anchors Area 45 as a core semantic node.

Area 44: The Motor-Articulatory Interface
Area 44 reveals a fundamentally different architecture. The partial correlation model shows no meaningful integration with the parietal or visual motion networks. Regions such as MT, MST, and PFop demonstrate near-zero or negative shared variance (). Instead, Area 44 strictly isolates its functional coupling to the anterior ventral insula (AVI: left , ; right , ), the premotor language node 55b (left , ), and the IFJa (left , ; right , ).

4.6 Behavioral Prediction

To check whether the resting-state networks identified in sections 4.2–4.5 are functionally meaningful, we tested if individual differences in RSFC could predict how well participants performed on behavioral tasks designed to target the auditory ‘what’ and ‘where’ pathways.

4.6.1 “What” Stream: Predicting Semantic Processing Speed

The main test of the ventral ‘what’ stream hypothesis used the Language Task Story median reaction time (Median RT). The ventral ROI model significantly predicted story Median RT (, ), with area A5 and STGa as the strongest positive predictors — meaning that stronger resting-state coupling to area A5 and STGa was associated with slower response times. The dorsal ROI model did not reach significance (, n.s.). This provides a clean dissociation: the semantic ‘what’ network predicts language comprehension speed, while the spatial ‘where’ network does not.

[PLOT A — Number of edges per area, Language Task Median RT, ventral model, part371, K=371]
What to see: Area 45 should dominate the positive edge count, with few or no other areas contributing substantially. This visualizes why Area 45 drives the prediction and anchors the IFJa–Area 45 coupling established in section 4.3.2 as the behavioral correlate of semantic processing speed.

The story accuracy (Acc) results told a different story. The full network predicted story Acc (, ), and both the ventral (, ) and dorsal (, ) submodels reached significance. Interestingly, area 45 appeared as a negative predictor of accuracy — stronger Broca’s area coupling was associated with lower accuracy. Given the simple vocabulary of the Aesop fable stories, this might point towards an over-engagement of deep semantic and syntactic machinery that actually disrupts fluent comprehension rather than helping it. An alternative explanation cannot be ruled out, though: because accuracy was near ceiling, the remaining variance may simply reflect attentional lapses or button errors, which would make the negative Area 45 contribution more likely a product of noise-fitting than a real effect. The dorsal model’s contribution, driven by areas 43, 7PC, and PF, likely reflects general attentional demand rather than stream-specific semantic processing. Reaction time is therefore the more meaningful and stream-specific measure for the ‘what’ stream hypothesis.

Story accuracy did not replicate in the larger full812 sample (n.s.) — consistent with the ceiling effect interpretation, since without matched behavioral data for each subject, there is simply not enough variance to predict. The Median RT effect, however, held up in the full812 all-ROI model with cross-validation (, ), showing that the semantic speed signal is robust enough to survive even when subject-level behavioral data is incomplete.

4.6.2 “Where” Stream: Predicting Visuo-Spatial Working Memory

We tested the dorsal ‘where’ stream hypothesis against Working Memory Task Place accuracy (Place Acc) and reaction time (Place RT). In the global model (all ROIs, part371, ), Place Acc was significantly predicted (, ), with SCEF (supplementary and cingulate eye field) and the anterior ventral insula (AVI) as the strongest positive predictors.

[PLOT B — Connectivity mask, positive predictive edges, WM Place Acc, full network model, part371, K=371]
What to see: SCEF should appear as the dominant node with the highest number of positive predictive edges. AVI may appear as a secondary node. FEF should be absent or show minimal edges — its absence from the positive predictors is notable given that SCEF sits anatomically adjacent to it in medial dorsal frontal cortex.

When we looked at stream-specific models, however, the picture reversed in a striking way. The dorsal ROI subset predicted Place Acc with a negative cross-validated R (, ), with superior parietal areas 7AL and 7Am contributing positively in an otherwise inverse model — a result that is hard to reconcile with a straightforward dorsal ‘where’ stream contribution. By contrast, the ventral ROI model predicted Place Acc significantly and positively (, ), driven by area 47l (lateral orbitofrontal/inferior frontal cortex). This suggests that performance on the visuo-spatial working memory task is better explained by ventral semantic network connectivity than by dorsal spatial network connectivity.

For Place RT, the dorsal model produced an artefactual result () caused by matrix rank collapse under partial correlation with too few ROIs — this result is excluded from interpretation (see section 5.7). The ventral model yielded a marginally significant prediction (, ), with area STSda contributing positively and IFJa negatively. The negative contribution of IFJa — stronger IFJa connectivity associated with faster Place RT — is consistent with the semantic hijacking interpretation above. If participants solve the Place task via object recognition rather than spatial navigation, then IFJa engagement reflects its role as a semantic controller (Soyuhos, O., & Baldauf, D. (2023), Bedini, M., & Baldauf, D. (2021)), not a domain-general executive response. Notably, Bedini & Baldauf (2021) also report IFJa involvement in working memory, adding another layer of consistency to this finding — IFJa’s engagement in the Place task may reflect both its semantic control function and its role in maintaining object representations in working memory. This reading is reinforced by area 47l also driving Place Accuracy in the same ventral model: both predictors point toward a semantic processing strategy as the dominant mechanism. In other words, the fact that IFJa appears in a visuo-spatial task does not contradict the ventral stream specificity from section 4.6.1 — it instead provides further evidence that the task itself was solved semantically rather than spatially. The full-network model (, ) identified MT and TA2 as the strongest negative predictors of Place RT — stronger connectivity to visual motion and auditory association cortex was linked to faster responses — while PSL emerged as the leading positive predictor, suggesting that stronger perisylvian language area coupling was associated with slower Place RT.

Taken together, these results suggest that the 2-Back Place task was predominantly solved via ventral object-recognition strategies rather than dorsal spatial navigation, consistent with the methodological constraint noted in section 3.4.2. The FEF did not emerge as a significant predictor in either the full-network or the stream-specific models — a null result consistent with both the task’s visual nature and the semantic hijacking interpretation above.

The dominant predictor in the full-network model was instead SCEF (supplementary and cingulate eye field). SCEF is located in medial frontal cortex as part of the SMA/cingulate motor complex and is classified within the dorsal network in section 3.2. Despite its vocalization selectivity per Dureux (2024) — it responds almost exclusively to vocalizations, suggesting cingulate attention monitoring of biologically relevant sounds — its anatomical position firmly places it in the dorsal attentional architecture adjacent to the FEF. Its emergence as the top predictor for a visuo-spatial working memory task is therefore consistent with a dorsal attentional contribution, though the modest effect size () and the failure of the isolated dorsal submodel to reach a positive result suggest that SCEF’s role operates through distributed network interactions rather than a dedicated spatial processing pathway.

4.6.3 Control Paradigm: Acoustic Signal Filtering (Noise Comparison)

The NIH Toolbox Noise Comparison score served as a control for low-level acoustic signal filtering. In contrast to the language comprehension results, the ventral ROI model failed to predict Noise Comparison performance (, n.s.). This means that the RSFC of the IFJa-anchored semantic network does not generalise to acoustic noise exclusion.

Significant predictions came instead from the full-network model (, ) and the dorsal ROI model (, ). In the full-network model, area PFcm (inferior parietal cortex, medial PF complex) was the dominant positive predictor, with FOP3 as the main negative predictor. The dorsal submodel showed a different pattern: OP4 (parietal operculum, area 4) was the dominant positive predictor, while FOP3, 7AL, and PBelt contributed negatively. Across both models, the leading predictors are opercular and parietal rather than prefrontal — which dissociates acoustic filtering from the IFJa-anchored semantic control network.

[PLOT C — R-value overview, all tasks × ventral/dorsal/full models, part371, K=371]
What to see: Three task groups (Language RT, WM Place Acc, Noise Comp), each with three bars (ventral, dorsal, full). The pattern should show: (1) ventral dominates for Language RT, dorsal flat — the ‘what’-stream dissociation; (2) ventral positive, dorsal negative for Place Acc — the semantic hijacking pattern; (3) ventral near zero, dorsal and full positive for Noise Comp — the double dissociation with acoustic filtering. Significant bars marked. This figure summarizes the entire argument of section 4.6 in a single view.

This double dissociation — ventral specificity for semantic RT, parieto-opercular dominance for acoustic filtering — suggests a hierarchical organization: the ‘cocktail party’ problem of separating physical signals is solved at an earlier, lower-level auditory processing stage before the IFJa-anchored prefrontal network steps in for meaning extraction.


5.0 Discussion

5.1 Resting-State Evidence for a Supramodal Prefrontal Architecture

Our central objective of this study was to examine whether the prefrontal hubs known to govern top-down attentional control in the visual domain extend their functional architecture into the auditory system. Our resting-state connectivity patterns are consistent with this hypothesis. Applying single-seed partial correlation analyses, we observe a functional dissociation between the FEF and the IFJa: the FEF shows preferential resting-state coupling with auditory-spatial and motion-sensitive cortical regions, whereas the IFJa couples selectively with temporal-semantic and language-related areas. This pattern supports our primary hypothesis (Section 1.3) that the auditory where-stream connects preferentially to FEF and the what-stream to IFJa. The observed dissociation is analogous to the spatial–non-spatial segregation previously established for these regions in the visual domain by Bedini & Baldauf (2021), who characterized the FEF as a spatial attention hub embedded in the dorsal attention system and the IFJa as anchoring for non-spatial, object-level processing. That our auditory connectivity profiles mirror this division suggests a prefrontal organization that may not be modality-specific but supramodal in its logic.

5.1.1 The FEF as an Auditory-Spatial Controller

The partial correlation profile of the FEF is consistent with its established role as a spatial attention controller, now extended to the auditory domain. Rather than projecting to the superior parietal lobule regions typically associated with visuospatial attention (7PC, 7Am, 7AL) — connections that vanish when shared ROI variance is partialled out — the FEF couples selectively with the inferior parietal lobule, specifically PF (left ; right ) and PFcm, as well as with the motion-sensitive area MST (bilateral, ) and multimodal convergence zones including STV and TPOJ1. Crucially, the FEF shows no significant effect-size connectivity with any temporal regions of the ventral “what” stream (e.g., TE1a, TA2, TGd), suggesting a topographically clean dissociation.

This pattern aligns with the functional logic described by Rauschecker & Scott (2009), who demonstrated that the posterior-dorsal auditory stream links posterior superior temporal cortex and parietal areas for spatial processing, with activation in regions adjacent to MT/MST specifically for auditory motion. Our resting-state data extend this framework to the prefrontal level by showing that the FEF — a node absent from classical auditory stream models — shows functional coupling with these motion-tracking areas, suggesting it may exert top-down control over auditory spatial orienting via the inferior parietal lobule. This stands in contrast to the superior parietal lobule route that characterizes visuospatial FEF connectivity Bedini & Baldauf (2021), a shift that may reflect the inherently multimodal nature of auditory spatial processing: auditory spatial localization operates within a distributed, multisensory reference frame rather than a retinotopic, egocentric one Rauschecker & Scott (2009).

5.1.2 The IFJa as a Semantic-Auditory Controller

The partial correlation profile of the IFJa reveals a complementary picture. After controlling for shared ROI variance, IFJa couples robustly with superior temporal sulcus regions central to auditory object identity processing (STSdp: left , right ; STSda: left , right ), with the Broca’s area subregions BA44 (bilateral ) and BA45 (bilateral ), and with early auditory association areas A4 and A5, while showing no substantial coupling with parietal spatial areas. In direct contrast to the FEF, the IFJa’s partial connectivity profile is predominantly language- and identity-oriented, and spatially disjoint from the parietal networks that define the dorsal where-stream.

This selectivity is consistent with the IFJa’s network membership and functional characterization. Per the resting-state fMRI parcellation of Ji et al. (2019, as reviewed in Bedini & Baldauf (2021)), the IFJa belongs to the language network, while the FEF is assigned to the cingulo-opercular network in that same partition — a classification our data directly support for the IFJa: its strongest auditory couplings are with the STS and Broca’s area. The concurrent coupling with early auditory association areas A4 and A5 suggests that IFJa may exert top-down modulation at an early stage of auditory feature extraction, biasing the processing of spectrotemporal features relevant to auditory object identity. The coupling with Broca’s area suggests coordination with the language working memory system, consistent with the IFJa’s proposed role in feature- and object-based attention encoding Bedini & Baldauf (2021). This pattern further resonates with the broader finding that top-down prefrontal control over auditory cortex is implemented via anticipatory alpha oscillations in object-based attention tasks De Vries & Baldauf (2021).

5.1.3 A Supramodal Prefrontal Architecture

Taken together, the complementary dissociation between FEF and IFJa connectivity profiles is consistent with a supramodal organization of prefrontal attentional control. The connectivity patterns suggest that the brain may not deploy dedicated, modality-specific attention controllers for each sensory stream; instead, the same two prefrontal hubs — FEF for spatial and IFJa for non-spatial control — appear to couple preferentially with corresponding auditory cortical regions in ways that parallel their known functional roles in the visual domain.

This is compatible with the view that the lateral prefrontal cortex organizes top-down attention according to a spatial/non-spatial axis that cuts across sensory modalities. Such a supramodal architecture would be computationally efficient — allowing a single prefrontal control system to flexibly coordinate attention across vision, audition, and potentially other modalities — and would be consistent with the evolutionary argument that visuospatial attention systems, which are phylogenetically older, are preserved in primates and subsequently recruited for the control of other sensory domains Bedini & Baldauf (2021). Whether these resting-state coupling patterns reflect genuine top-down control signals remains to be established with directed connectivity methods.

5.2 Re-evaluating the auditory where stream

As Hickock und Poeppel proposed a How pathway which might be the language oatheway rolls talked about

1. Neighboring Areas

The Resulkts frm 4.5 show another dissoctation within the where stream where FEF and 55b share the labour of this stream.

  • this could

2. Subheading


6.0 References

Ahveninen, J., Jääskeläinen, I. P., Raij, T., Bonmassar, G., Devore, S., Hämäläinen, M., Levänen, S., Lin, F.-H., Sams, M., Shinn-Cunningham, B. G., Witzel, T., & Belliveau, J. W. (2006). Task-modulated “what” and “where” pathways in human auditory cortex. Proceedings of the National Academy of Sciences, 103(39), 14608–14613. https://doi.org/10.1073/pnas.0510480103

Baker, C. M. [et al.] (2018). [CITATION NEEDED — aus Zotero ergänzen]

Bedini, M., & Baldauf, D. (2021). Structure, function and connectivity fingerprints of the frontal eye field versus the inferior frontal junction: A comprehensive comparison. European Journal of Neuroscience, 54(4), 5462–5506. https://doi.org/10.1111/ejn.15393

Bedini, M., Olivetti, E., Avesani, P., & Baldauf, D. (2023). Accurate localization and coactivation profiles of the frontal eye field and inferior frontal junction: An ALE and MACM fMRI meta-analysis. Brain Structure and Function, 228(3–4), 997–1017. https://doi.org/10.1007/s00429-023-02641-y

Bedini, M., Olivetti, E., Avesani, P., & Baldauf, D. (n.d.). Surface-based probabilistic tractography uncovers segregated white matter pathways underlying spatial and non-spatial control [Manuscript in preparation].

Benjamini, Y., & Hochberg, Y. (1995). Controlling the false discovery rate: A practical and powerful approach to multiple testing. Journal of the Royal Statistical Society Series B: Statistical Methodology, 57(1), 289–300. https://doi.org/10.1111/j.2517-6161.1995.tb02031.x

Biswal, B., Zerrin Yetkin, F., Haughton, V. M., & Hyde, J. S. (1995). Functional connectivity in the motor cortex of resting human brain using echo-planar MRI. Magnetic Resonance in Medicine, 34(4), 537–541. https://doi.org/10.1002/mrm.1910340409

Cohen, Y. E., Bennur, S., Christison-Lagay, K., Gifford, A. M., & Tsunada, J. (2016). Functional organization of the ventral auditory pathway. In P. Van Dijk, D. Başkent, E. Gaudrain, E. De Kleine, A. Wagner, & C. Lanting (Eds.), Physiology, Psychoacoustics and Cognition in Normal and Impaired Hearing (Vol. 894, pp. 381–388). Springer. https://doi.org/10.1007/978-3-319-25474-6_40

De Vries, I. E. J., Marinato, G., & Baldauf, D. (2021). Decoding object-based auditory attention from source-reconstructed MEG alpha oscillations. The Journal of Neuroscience, 41(41), 8603–8617. https://doi.org/10.1523/JNEUROSCI.0583-21.2021

Dureux, A. [et al.] (2024). [CITATION NEEDED — aus Zotero ergänzen]

Friederici, A. D. (2011). The brain basis of language processing: From structure to function. Physiological Reviews, 91(4), 1357–1392. https://doi.org/10.1152/physrev.00006.2011

Friston, K. J. (1994). Functional and effective connectivity in neuroimaging: A synthesis. Human Brain Mapping, 2(1–2), 56–78. https://doi.org/10.1002/hbm.460020107

Frühholz, S., Gschwind, M., & Grandjean, D. (2015). Bilateral dorsal and ventral fiber pathways for the processing of affective prosody identified by probabilistic fiber tracking. NeuroImage, 109, 27–34. https://doi.org/10.1016/j.neuroimage.2015.01.016

Glasser, M. F., Coalson, T. S., Robinson, E. C., Hacker, C. D., Harwell, J., Yacoub, E., Ugurbil, K., Andersson, J., Beckmann, C. F., Jenkinson, M., Smith, S. M., & Van Essen, D. C. (2016). A multi-modal parcellation of human cerebral cortex. Nature, 536(7615), 171–178. https://doi.org/10.1038/nature18933

Goodale, M. A., & Milner, A. D. (1992). Separate visual pathways for perception and action. Trends in Neurosciences, 15(1), 20–25. https://doi.org/10.1016/0166-2236(92)90344-8

Griffiths, T. D., Rees, G., Rees, A., Green, G. G. R., Witton, C., Rowe, D., Büchel, C., Turner, R., & Frackowiak, R. S. J. (1998). Right parietal cortex is involved in the perception of sound movement in humans. Nature Neuroscience, 1(1), 74–79. https://doi.org/10.1038/276

Hickok, G., & Poeppel, D. (2004). Dorsal and ventral streams: A framework for understanding aspects of the functional anatomy of language. Cognition, 92(1–2), 67–99. https://doi.org/10.1016/j.cognition.2003.10.011

Hickok, G., & Poeppel, D. (2007). The cortical organization of speech processing. Nature Reviews Neuroscience, 8(5), 393–402. https://doi.org/10.1038/nrn2113

Honey, C. J., Sporns, O., Cammoun, L., Gigandet, X., Thiran, J. P., Meuli, R., & Hagmann, P. (2009). Predicting human resting-state functional connectivity from structural connectivity. Proceedings of the National Academy of Sciences, 106(6), 2035–2040. https://doi.org/10.1073/pnas.0811168106

Ji, J. L., Spronk, M., Kulkarni, K., Repovš, G., Anticevic, A., & Cole, M. W. (2019). Mapping the human brain’s cortical-subcortical functional network organization. NeuroImage, 185, 35–57. https://doi.org/10.1016/j.neuroimage.2018.10.006

Marrelec, G., Krainik, A., Duffau, H., Pélégrini-Issac, M., Lehéricy, S., Doyon, J., & Benali, H. (2006). Partial correlation for functional brain interactivity investigation in functional MRI. NeuroImage, 32(1), 228–237. https://doi.org/10.1016/j.neuroimage.2005.12.057

Rauschecker, J. P. (2011). An expanded role for the dorsal auditory pathway in sensorimotor control and integration. Hearing Research, 271(1–2), 16–25. https://doi.org/10.1016/j.heares.2010.09.001

Rauschecker, J. P., & Afsahi, R. K. (2023). Anatomy of the auditory cortex then and now. Journal of Comparative Neurology, 531(18), 1883–1892. https://doi.org/10.1002/cne.25560

Rauschecker, J. P., & Scott, S. K. (2009). Maps and streams in the auditory cortex: Nonhuman primates illuminate human speech processing. Nature Neuroscience, 12(6), 718–724. https://doi.org/10.1038/nn.2331

Rolls, E. T., Deco, G., Huang, C.-C., & Feng, J. (2022). The human language effective connectome. NeuroImage, 258, 119352. https://doi.org/10.1016/j.neuroimage.2022.119352

Rolls, E. T., Rauschecker, J. P., Deco, G., Huang, C.-C., & Feng, J. (2023). Auditory cortical connectivity in humans. Cerebral Cortex, 33(10), 6207–6227. https://doi.org/10.1093/cercor/bhac496

Romanski, L. M., Tian, B., Fritz, J., Mishkin, M., Goldman-Rakic, P. S., & Rauschecker, J. P. (1999). Dual streams of auditory afferents target multiple domains in the primate prefrontal cortex. Nature Neuroscience, 2(12), 1131–1136. https://doi.org/10.1038/16056

Salmi, J., Rinne, T., Koistinen, S., Salonen, O., & Alho, K. (2009). Brain networks of bottom-up triggered and top-down controlled shifting of auditory attention. Brain Research, 1286, 155–164. https://doi.org/10.1016/j.brainres.2009.06.083

Scott, B. H., Saleem, K. S., Kikuchi, Y., Fukushima, M., Mishkin, M., & Saunders, R. C. (2017). Thalamic connections of the core auditory cortex and rostral supratemporal plane in the macaque monkey. Journal of Comparative Neurology, 525(16), 3488–3513. https://doi.org/10.1002/cne.24283

Smith, S. M., Miller, K. L., Salimi-Khorshidi, G., Webster, M., Beckmann, C. F., Nichols, T. E., Ramsey, J. D., & Woolrich, M. W. (2011). Network modelling methods for FMRI. NeuroImage, 54(2), 875–891. https://doi.org/10.1016/j.neuroimage.2010.08.063

Soyuhos, O., & Baldauf, D. (2023). Functional connectivity fingerprints of the frontal eye field and inferior frontal junction suggest spatial versus nonspatial processing in the prefrontal cortex. European Journal of Neuroscience, 57(7), 1114–1140. https://doi.org/10.1111/ejn.15936

Ungerleider, L. G., & Mishkin, M. (1982). Two cortical visual systems. In D. J. Ingle, M. A. Goodale, & R. J. W. Mansfield (Eds.), Analysis of visual behavior (pp. 549–586). MIT Press.

Veniero, D., Gross, J., Morand, S., Duecker, F., Sack, A. T., & Thut, G. (2021). Top-down control of visual cortex by the frontal eye fields through oscillatory realignment. Nature Communications, 12(1), 1757. https://doi.org/10.1038/s41467-021-21979-7

Wernicke, C. (1874). Der aphasische Symptomenkomplex: Eine psychologische Studie auf anatomischer Basis. Cohn & Weigert.

Wu-Minn HCP Consortium. (2018). WU-Minn HCP 1200 Subjects Data Release: Reference Manual. Human Connectome Project. https://www.humanconnectome.org/storage/app/media/documentation/s1200/HCP_S1200_Release_Reference_Manual.pdf


7.0 Appendix

7.1 Glasser Atlas (2016)

Here is the parcellation created by Glasser

2. Subheading


8.0 Declaration

1. Subheading

Hier schreiben…

2. Subheading


9.0 Abbreviations

1. Brain Region Abbreviations

AbbreviationFull Name
43Area 43
44 / BA44Area 44, Pars Opercularis (Broca’s area)
45 / BA45Area 45, Pars Triangularis (Broca’s area)
47lArea 47 Lateral
55bArea 55b
7ALArea 7 Anterior Lateral
7AmArea 7 Anterior Medial
7PCArea 7 Posterior Capsular
A1Primary Auditory Cortex (Area A1)
A4Auditory Area 4
A5Auditory Area 5
AVIAnterior Ventral Insula
BABrodmann Area
FEFFrontal Eye Field
FOP1–3Frontal Operculum 1, 2, 3
IFJInferior Frontal Junction
IFJaanterior Inferior Frontal Junction
IFJpposterior Inferior Frontal Junction
IFSInferior Frontal Sulcus
IFSpInferior Frontal Sulcus posterior
iPCSinferior precentral sulcus
IPLInferior Parietal Lobule
MSTMedial Superior Temporal Area
MTMiddle Temporal Area
OP1–4Opercular Areas 1–4
PBeltParabelt Complex
PFArea PF (Inferior Parietal)
PFCPrefrontal Cortex
PFcmArea PF Complex Medial
PFopArea PF Opercular
PGiArea PGi (Inferior Parietal)
PMCPremotor Cortex
PSLPerisylvian Language Area
SCEFSupplementary and Cingulate Eye Field
SPLSuperior Parietal Lobule
SptSylvian parieto-temporal area
SFSSuperior Frontal Sulcus
sPCSsuperior precentral sulcus
STSSuperior Temporal Sulcus
STSdaSuperior Temporal Sulcus, dorsal anterior
STSdpSuperior Temporal Sulcus, dorsal posterior
STSvaSuperior Temporal Sulcus, ventral anterior
STSvpSuperior Temporal Sulcus, ventral posterior
STGSuperior Temporal Gyrus
STGaSuperior Temporal Gyrus, anterior
STVSuperior Temporal Visual Area
TA2Area TA2
TE1aArea TE1 Anterior
TGdTemporal Gyrus dorsal (Temporal Pole)
TGvTemporal Gyrus ventral (Temporal Pole)
TPOJ1Temporoparietal Occipital Junction Area 1
V1Primary Visual Cortex

2. Other Abbreviations

AbbreviationFull Name
BOLDBlood Oxygen Level Dependent
CIFTIConnectivity Informatics Technology Initiative (HCP file format)
DANDorsal Attention Network
ECEffective Connectivity
EPIEcho-Planar Imaging
RSFCResting-State Functional Connectivity
FDRFalse Discovery Rate
fMRIfunctional Magnetic Resonance Imaging
FPNFrontoparietal Network
HCPHuman Connectome Project
HCP-MMP1HCP Multimodal Parcellation (version 1.0)
ICAIndependent Component Analysis
ICA-FIXICA-based X-noiseifier
LHleft hemisphere
MDMultiple Demand (system)
MEGMagnetoencephalography
RHright hemisphere
ROIRegion of Interest
rs-fMRIresting-state functional Magnetic Resonance Imaging
SLFSuperior Longitudinal Fasciculus
T1wT1-weighted (MRI)
T2wT2-weighted (MRI)
TEEcho Time
TRRepetition Time
VANVentral Attention Network