Spoken language comprehension relies on both the identification of individual words and the expectations arising from contextual information. A distributed fronto-temporal network is known to facilitate the mapping of speech sounds onto corresponding meanings. However, how prior expectations influence this efficient mapping at the neuroanatomical level, especially for individual words, remains unclear. Using functional magnetic resonance imaging, we addressed this question in the framework of the dual-stream model by investigating both the neural substrates and their mutual functional and effective connectivity. Our results revealed how this ubiquitous sound-to-meaning mapping in daily communication is achieved in a predictive manner.