Our lab uses quantitative, computational methods to try to understand how the human brain processes the natural world. In particular, we are focused on understanding how the meaning of language is represented in the brain.
Using fMRI, we record human brain responses while people listen to speech in the form of stories or podcasts. Then we build encoding models that predict those responses based on the audio and transcript of the stories. The best encoding models today use neural network language models to extract meaningful information from the stories. Our work uses encoding models to map how language is represented across the brain [Jain et al., 2018, 2020; Antonello et al., 2021], investigates why neural network language models are so effective [Antonello & Huth, 2023], and shows that we can even decode language from fMRI [Tang et al., 2023].
The datasets we collect are shared freely and we encourage you to use them [LeBel et al., 2022].
We also share code that demonstrates how to use these datasets and tutorials on encoding models for language.
Tang, Jerry and Huth, Alexander G (2025). Semantic language decoding across participants and stimulus modalities. Current Biology. doi: 10.1016/j.cub.2025.01.024
Abdel-Ghaffar, Samy A and Huth, Alexander G and Lescroart, Mark D and Stansbury, Dustin and Gallant, Jack L and Bishop, Sonia J (2024). Occipital-temporal cortical tuning to semantic and affective features of natural images predicts associated behavioral responses. Nature communications. doi: 10.1016/j.cub.2025.01.024
Vinamra Benara and Chandan Singh and John Xavier Morris and Richard Antonello and Ion Stoica and Alexander Huth and Jianfeng Gao (2024). Crafting Interpretable Embeddings for Language Neuroscience by Asking LLMs Questions. Advances in Neural Information Processing Systems. (paper) (GitHub)