Table of Contents

Building a Collaborative Foundation for Non-Invasive Speech BCIs: The 2025 PNPL Competition

Exploring the motivation behind the 2025 PNPL Competition and how we're building towards non-invasive speech brain-computer interfaces through deep MEG datasets and collaborative research.

January 20, 2025•7 min read
By Gereon Elvers, Gilad Landau, Dulhan Jayalath, Francesco Mantegna, Oiwi Parker Jones
Brain-Computer InterfacesMEGSpeech DecodingLibriBrainCompetition

Note: This blog post was originally published on the LibriBrain Competition website and is reproduced here with permission.

The field of brain-computer interfaces (BCIs) has reached an exciting inflection point. Recent breakthroughs in invasive speech BCIs have achieved remarkable milestones—reported word-error rates below 5% and vocabularies exceeding 125,000 words. But these advances come with a fundamental limitation: they require brain surgery.

Brain surgery procedure
⚠️ Click to view surgical image
(Warning: contains blood and medical content)

Part of the surgery to implant invasive electrodes into the brain (from Wikipedia)

What if we could achieve similar breakthroughs without surgical intervention? This is the driving vision behind the 2025 PNPL Competition, and we're excited to share why we believe this competition represents a critical step towards that future.

The Ultimate Goal: Speech BCIs Without Surgery

Our ultimate goal is clear: developing speech brain-computer interfaces that don't require surgical implantation. The potential impact is profound—from restoring communication for individuals with paralysis or speech deficits like dysarthria, to enabling new forms of human-computer interaction that could benefit everyone.

However, non-invasive speech decoding faces significant challenges. While invasive BCIs leverage high signal-to-noise ratios from electrodes placed directly on or in the brain, non-invasive methods must contend with signals that are attenuated by the distance from brain and sensor. The gap between invasive and non-invasive performance has been substantial—most current non-invasive approaches show word-error rates approaching 100% (though see Jayalath et al.), whilst invasive systems report below 5%.

Addressing Field Limitations with LibriBrain

To tackle these challenges head-on, we've created LibriBrain: the largest within-subject MEG dataset recorded to date, containing over 50 hours of high-quality 306-channel neural recordings sampled at 250 Hz. This represents a 5Ă— increase over the next largest dataset and 25-50Ă— more data than typical MEG/EEG datasets.

LibriBrain dataset comparison

Figure reproduced from Ă–zdogan et al.

The dataset spans nearly the entire Sherlock Holmes canon across 7 books, recorded over 95 sessions from a single participant. This depth matters enormously—our baseline experiments demonstrate clear logarithmic scaling relationships between training data volume and decoding performance, consistent with scaling laws observed across machine learning domains.

Why does this matter? Empirical evidence shows that "deep data"—extensive recordings from the same individual—yields the largest gains in decoding performance. Previous datasets have been characterised as "broad but shallow," including many participants but only 1-2 hours per person. LibriBrain flips this paradigm, providing the depth needed to train powerful AI models that can compete with surgical alternatives. That said, as we continue to grow the dataset, we also plan in future to release data from additional subjects.

Competition Design: Building Strong Foundations

Why These Tasks?

We've chosen two foundational tasks for this competition: Speech Detection and Phoneme Classification. Whilst our ultimate goal is brain-to-text decoding, recent attempts at full sentence decoding from non-invasive signals have yielded near-chance performance.

Instead of jumping directly to the hardest problem, we're taking inspiration from the development of automatic speech recognition (ASR), which built strong foundations through phoneme-based approaches. Here's why these tasks make sense:

Speech Detection provides a binary classification for every temporal sample (250 per second), making it our most accessible task—maximum data points with the smallest number of classes, compared for example to word or even phoneme events. This entry-level challenge mirrors the critical speech detection component used in the first invasive speech BCI for paralysed individuals.

Speech detection visualization

Phoneme Classification offers several advantages over word-level tasks:

  • No out-of-vocabulary problems: Unlike word classification, every phoneme in test data has been seen during training
  • Better data efficiency: 1.5M phoneme examples across 39 classes versus 466k words across 16k+ classes
  • Strong theoretical foundation: Phonemes are the building blocks of speech

Phoneme classification visualization

Our baseline results show promising scaling behaviour which mirrors neural scaling laws observed across AI domains (e.g. Kaplan et al. and Antonello et al.). Both speech detection and phoneme classification performance improve logarithmically with the volume of training data:

Scaling results

Figure reproduced from Ă–zdogan et al. (x-axes are log scale)

Why MEG?

Magnetoencephalography (MEG) offers unique advantages that make it ideal for bridging the gap between invasive and practical non-invasive BCIs:

  • Direct measurement of neural activity with millisecond temporal resolution
  • Minimal signal distortion compared to EEG, as magnetic fields pass through tissue without degradation
  • Spatial precision of 5-10mm routinely, with 2mm achieved in some studies
  • No surgical risk, enabling safe data collection at unprecedented scale

Recent work by Défossez et al., d'Ascoli et al., and Jayalath et al. has demonstrated the potential of scaling non-invasive approaches with large datasets and modern AI methods. Whilst practical applications may eventually move towards more portable technologies like EEG, MEG provides the best current platform for proving that non-invasive speech decoding can compete with surgical alternatives.

MEG scanner MEG provides high-quality, non-invasive brain recordings (Oxford scanner used for LibriBrain)

Why Speech Comprehension?

Our competition focuses on decoding speech comprehension (listening) rather than speech production. This design choice reflects both practical and strategic considerations:

Methodological advantages: Speech comprehension allows us to develop scalable AI methods with precise event timing. We know exactly when each phoneme occurs, enabling reliable model training and evaluation. Comprehension tasks further avoid muscle artifacts introduced by speech production (e.g. speaking aloud) which reduce signal quality and confound non-invasive decoding.

Alignment with literature: This approach aligns with recent successful non-invasive decoding work, including influential papers by Tang et al., Défossez et al., and our own recent advances in non-invasive brain-to-text (e.g. Jayalath et al. and Jayalath et al.) that show promise for generalising from comprehension to production tasks.

Stepping stone to production: Whilst we ultimately want to decode attempted speech or inner speech, comprehension provides a crucial intermediate step for developing and validating methods that can later generalise to production tasks. This approach is supported by evidence of overlap in the brain between speech comprehension and production representations (e.g. Hickok and Poeppel and PulvermĂĽller et al.), and by recent evidence that decoding models trained on listening data can transfer to inner speech tasks (e.g. Tang et al.).

Driving Innovation and Building Community

This competition is designed to create an "ImageNet moment" for non-invasive neural decoding, inspired by how ImageNet revolutionised computer vision. We're providing not just data, but a complete ecosystem designed around the foundational tasks we've established:

  • Standardised evaluation metrics to enable fair comparisons
  • Easy-to-use Python library (pnpl) with PyTorch integration, designed to be as accessible as CIFAR-10
  • Two competition tracks: Standard (LibriBrain data only) and Extended (any training data)
  • Tutorial materials that run in free Google Colab environments
  • Active community support through Discord and dedicated resources
  • Public leaderboards for real-time progress tracking

The dual-track structure ensures both algorithmic innovation (Standard track) and exploration of what's possible with unlimited compute resources (Extended track). At least ÂŁ10,000 in prizes will reward the most innovative approaches.

Looking Forward

This is just the beginning. The hope is that the PNPL Competition will become an annual event, tackling a progression of increasingly challenging tasks as our datasets grow and our methods and insights improve. Think of it as a curriculum for the field—building from strong foundations towards the ultimate goal of robust, non-invasive speech BCIs.

We'll be releasing a series of blog posts throughout the competition to dive deeper into technical aspects, share insights from the community, and highlight innovative approaches. Whether you're a seasoned researcher or new to the field, we've designed this competition to be accessible and impactful.

The future of speech BCIs doesn't have to require surgery—let's work together to make this future a reality


Ready to participate? Visit the rest of our competition website to get started with tutorials, download the data, and join our community Discord. The competition is officially running from 1st June 2025 - 30 September 2025 and will culminate in a session at NeurIPS in December 2025. As a record of progress, we expect the submission system and leaderboards to continue running afterward.


đź“„Cite this blog post

Use this citation format to reference this blog post in your research

@misc{pnpl_blog2025collaboration,
  title={Building a Collaborative Foundation for Non-Invasive Speech {BCIs}: The 2025 {PNPL} Competition},
  author={Elvers, Gereon and Landau, Gilad and Jayalath, Dulhan and Mantegna, Francesco and Parker Jones, Oiwi},
  year={2025},
  url={https://neural-processing-lab.github.io/2025-libribrain-competition/blog/building-collaborative-foundation-non-invasive-speech-bcis/},
  note={Blog post}
}

References

[1]

A high-performance neuroprosthesis for speech decoding and avatar control

Sean L. Metzger, Kaylo T. Littlejohn, Alexander B. Silva, David A. Moses, Margaret P. Seaton, Ran Wang, Maximilian E. Dougherty, Jennifer L. Collinger, Michael L. Boninger, Edward F. Chang

NatureVol. 620pp. 1031-1036(2023)
[2]

A high-performance speech neuroprosthesis

Francis R. Willett, Erin M. Kunz, Chaofei Fan, Donald T. Avansino, Guy H. Wilson, Eun Young Choi, Foram Kamdar, Matthew F. Glasser, Leigh R. Hochberg, Shaul Druckmann, Krishna V. Shenoy, Jaimie M. Henderson

NatureVol. 620pp. 1031-1036(2023)
[3]

Unlocking non-invasive brain-to-text

Dulhan Jayalath, Gilad Landau, Oiwi Parker Jones

arXiv preprint(2025)
[4]

LibriBrain: Over 50 hours of within-subject MEG to improve speech decoding methods at scale

Miran Ă–zdogan, Gilad Landau, Gereon Elvers, Dulhan Jayalath, Pranav Somaiya, Francesco Mantegna, Mark Woolrich, Oiwi Parker Jones

arXiv preprint(2025)
[5]

Neuroprosthesis for decoding speech in a paralyzed person with anarthria

David A. Moses, Sean L. Metzger, Jessie R. Liu, Gopala K. Anumanchipalli, Joseph G. Makin, Pengfei F. Sun, Josh Chartier, Maximilian E. Dougherty, Patricia M. Liu, Gary M. Abrams, Adelyn Tu-Chan, Karunesh Ganguly, Edward F. Chang

New England Journal of MedicineVol. 385pp. 217-227(2021)
[6]

Scaling laws for neural language models

Jared Kaplan, Sam McCandlish, Tom Henighan, Tom B. Brown, Benjamin Chess, Rewon Child, Scott Gray, Alec Radford, Jeffrey Wu, Dario Amodei

arXiv preprint(2020)
[7]

Decoding speech perception from non-invasive brain recordings

Alexandre Défossez, Charlotte Caucheteux, Jérémy Rapin, Ori Kabeli, Jean-Rémi King

Nature Machine IntelligenceVol. 5pp. 1097-1107(2023)
[8]

Semantic reconstruction of continuous language from non-invasive brain recordings

Jerry Tang, Amanda LeBel, Shailee Jain, Alexander G. Huth

Nature NeuroscienceVol. 26pp. 858-866(2023)
[9]

The 2025 PNPL competition: Speech detection and phoneme classification in the LibriBrain dataset

Gilad Landau, Miran Ă–zdogan, Gereon Elvers, Francesco Mantegna, Pranav Somaiya, Dulhan Jayalath, Lukas Kurth, Taewon Kwon, Brendan Shillingford, Gabriel Farquhar, Michael Jiang, Karim Jerbi, Hichem Abdelhedi, Yorguin Mantilla Ramos, Caglar Gulcehre, Mark Woolrich, Natalie Voets, Oiwi Parker Jones

NeurIPS Competition Track(2025)