Unknown

Dataset Information

0

Simple Acoustic Features Can Explain Phoneme-Based Predictions of Cortical Responses to Speech.


ABSTRACT: When we listen to speech, we have to make sense of a waveform of sound pressure. Hierarchical models of speech perception assume that, to extract semantic meaning, the signal is transformed into unknown, intermediate neuronal representations. Traditionally, studies of such intermediate representations are guided by linguistically defined concepts, such as phonemes. Here, we argue that in order to arrive at an unbiased understanding of the neuronal responses to speech, we should focus instead on representations obtained directly from the stimulus. We illustrate our view with a data-driven, information theoretic analysis of a dataset of 24 young, healthy humans who listened to a 1 h narrative while their magnetoencephalogram (MEG) was recorded. We find that two recent results, the improved performance of an encoding model in which annotated linguistic and acoustic features were combined and the decoding of phoneme subgroups from phoneme-locked responses, can be explained by an encoding model that is based entirely on acoustic features. These acoustic features capitalize on acoustic edges and outperform Gabor-filtered spectrograms, which can explicitly describe the spectrotemporal characteristics of individual phonemes. By replicating our results in publicly available electroencephalography (EEG) data, we conclude that models of brain responses based on linguistic features can serve as excellent benchmarks. However, we believe that in order to further our understanding of human cortical responses to speech, we should also explore low-level and parsimonious explanations for apparent high-level phenomena.

SUBMITTER: Daube C 

PROVIDER: S-EPMC6584359 | biostudies-literature | 2019 Jun

REPOSITORIES: biostudies-literature

altmetric image

Publications

Simple Acoustic Features Can Explain Phoneme-Based Predictions of Cortical Responses to Speech.

Daube Christoph C   Ince Robin A A RAA   Gross Joachim J  

Current biology : CB 20190523 12


When we listen to speech, we have to make sense of a waveform of sound pressure. Hierarchical models of speech perception assume that, to extract semantic meaning, the signal is transformed into unknown, intermediate neuronal representations. Traditionally, studies of such intermediate representations are guided by linguistically defined concepts, such as phonemes. Here, we argue that in order to arrive at an unbiased understanding of the neuronal responses to speech, we should focus instead on  ...[more]

Similar Datasets

| S-EPMC5031534 | biostudies-literature
| S-EPMC11233028 | biostudies-literature
| S-EPMC8236853 | biostudies-literature
| S-EPMC7546411 | biostudies-literature
| S-EPMC9767914 | biostudies-literature
| S-EPMC11296674 | biostudies-literature
| S-EPMC4380160 | biostudies-literature
| S-EPMC10774448 | biostudies-literature
| S-EPMC9308078 | biostudies-literature
| S-EPMC6429566 | biostudies-literature