Back to Explorer
Research PaperResearchia:202602.23066[Bio-AI Interfaces > Neuroscience]

MEG-to-MEG Transfer Learning and Cross-Task Speech/Silence Detection with Limited Data

Xabier de Zuazo

Abstract

Data-efficient neural decoding is a central challenge for speech brain-computer interfaces. We present the first demonstration of transfer learning and cross-task decoding for MEG-based speech models spanning perception and production. We pre-train a Conformer-based model on 50 hours of single-subject listening data and fine-tune on just 5 minutes per subject across 18 participants. Transfer learning yields consistent improvements, with in-task accuracy gains of 1-4% and larger cross-task gains of up to 5-6%. Not only does pre-training improve performance within each task, but it also enables reliable cross-task decoding between perception and production. Critically, models trained on speech production decode passive listening above chance, confirming that learned representations reflect shared neural processes rather than task-specific motor activity.


Source: arXiv:2602.18253v1 - http://arxiv.org/abs/2602.18253v1 PDF: https://arxiv.org/pdf/2602.18253v1 Original Link: http://arxiv.org/abs/2602.18253v1

Submission:2/23/2026
Comments:0 comments
Subjects:Neuroscience; Bio-AI Interfaces
Original Source:
View Original PDF
arXiv: This paper is hosted on arXiv, an open-access repository
Was this helpful?

Discussion (0)

Please sign in to join the discussion.

No comments yet. Be the first to share your thoughts!

MEG-to-MEG Transfer Learning and Cross-Task Speech/Silence Detection with Limited Data | Researchia | Researchia