Decoding EEG signals of visual brain representations with a CLIP based knowledge distillation

ICLR 2024 Workshop TS4H Submission5 Authors

Published: 08 Mar 2024, Last Modified: 25 Mar 2024TS4H PosterEveryoneRevisionsBibTeXCC BY 4.0
Keywords: decoding, eeg, brain computer interface, image reconstruction
TL;DR: We developed a novel student teacher approach to decode EEG signals of visual content into images
Abstract: Decoding visual representations from human brain activity has emerged as a thriving research domain, particularly in the context of brain-computer interfaces. Our study presents an innovative method that employs to classify and reconstruct images from the ImageNet dataset using only electroencephalography (EEG) data from subjects that have viewed the images themselves (i.e. "brain decoding"). We analyzed EEG recordings from 6 participants, each exposed to 50 images spanning 40 unique semantic categories. These EEG readings were converted into spectrograms, which were then used to train a convolutional neural network (CNN), integrated with a knowledge distillation procedure based on a pre-trained Contrastive Language-Image Pre-Training (CLIP)-based image classification teacher network. This strategy allowed our model to attain a top-5 accuracy of 80\%, significantly outperforming a standard CNN and various RNN-based benchmarks. Additionally, we incorporated an image reconstruction mechanism based on pre-trained latent diffusion models, which allowed us to generate an estimate of the images that had elicited EEG activity. Therefore, our architecture not only decodes images from neural activity but also offers a credible image reconstruction from EEG only, paving the way for, e.g., swift, individualized feedback experiments.
Submission Number: 5
Loading