decoding visual representations from human brain activity has emerged as a thriving research domain, particularly in the context of brain-computer interfaces. this study introduces a novel approach using a convolutional neural network (CNN) to classify images from the ImageNet dataset, leveraging electroencephalography (EEG) recordings. we collected EEG data from 6 subjects, each viewing 50 images across 40 distinct semantic categories. these EEG signals were transformed into spectrograms, serving as the input for training our CNN. a unique aspect of our model is the incorporation of knowledge distillation from a pre-trained image classification teacher network. this approach enabled our model to achieve a top-5 accuracy of 80%, notably surpassing a plain CNN baseline. furthermore, we integrated an image reconstruction pipeline founded on pre-trained latent diffusion models. this innovative concatenation not only decodes images from brain activity but also provides a plausible reconstruction, facilitating rapid and subject-specific feedback experiments. our work thus represents a significant advancement in the field, bridging the gap between neural signals and visual perception.

Ferrante, M., Boccato, T., Bargione, S., Toschi, N. (2023). Linking Brain Signals to Visual Concepts: CLIP based knowledge transfer for EEG Decoding and visual stimuli reconstruction, 31-32 [10.1109/IEEECONF58974.2023.10404307].

Linking Brain Signals to Visual Concepts: CLIP based knowledge transfer for EEG Decoding and visual stimuli reconstruction

Ferrante M.;Boccato T.;Toschi N.
2023-01-01

Abstract

decoding visual representations from human brain activity has emerged as a thriving research domain, particularly in the context of brain-computer interfaces. this study introduces a novel approach using a convolutional neural network (CNN) to classify images from the ImageNet dataset, leveraging electroencephalography (EEG) recordings. we collected EEG data from 6 subjects, each viewing 50 images across 40 distinct semantic categories. these EEG signals were transformed into spectrograms, serving as the input for training our CNN. a unique aspect of our model is the incorporation of knowledge distillation from a pre-trained image classification teacher network. this approach enabled our model to achieve a top-5 accuracy of 80%, notably surpassing a plain CNN baseline. furthermore, we integrated an image reconstruction pipeline founded on pre-trained latent diffusion models. this innovative concatenation not only decodes images from brain activity but also provides a plausible reconstruction, facilitating rapid and subject-specific feedback experiments. our work thus represents a significant advancement in the field, bridging the gap between neural signals and visual perception.
2023
Pubblicato
Rilevanza internazionale
Articolo
Esperti anonimi
Settore PHYS-06/A - Fisica per le scienze della vita, l'ambiente e i beni culturali
English
This work was supported by NEXTGENERATIONEU (NGEU) and funded by the Italian Ministry of University and Research (MUR), National Recovery and Resilience Plan (NRRP), project MNESYS (PE0000006)– A Multiscale integrated approach to the study of the nervous system in health and disease (DN. 1553 11.10.2022); by the MUR-PNRR M4C2I1.3 PE6 project PE00000019 Heal Italia; by the NATIONAL CENTRE FOR HPC, BIG DATA AND QUANTUM COMPUTING, within the spoke ”Multiscale Modeling and Engineering Applications”; the EXPERIENCE project (European Union’s Horizon 2020 Research and Innovation Programme under grant agreement No. 101017727); the CROSSBRAIN project (European Union’s European Innovation Council under grant agreement No. 101070908).
Ferrante, M., Boccato, T., Bargione, S., Toschi, N. (2023). Linking Brain Signals to Visual Concepts: CLIP based knowledge transfer for EEG Decoding and visual stimuli reconstruction, 31-32 [10.1109/IEEECONF58974.2023.10404307].
Ferrante, M; Boccato, T; Bargione, S; Toschi, N
Articolo su rivista
File in questo prodotto:
File Dimensione Formato  
Ferrante-2023-Linking_Brain_Signals_to_Visual.pdf

solo utenti autorizzati

Tipologia: Versione Editoriale (PDF)
Licenza: Copyright dell'editore
Dimensione 622.87 kB
Formato Adobe PDF
622.87 kB Adobe PDF   Visualizza/Apri   Richiedi una copia

I documenti in IRIS sono protetti da copyright e tutti i diritti sono riservati, salvo diversa indicazione.

Utilizza questo identificativo per citare o creare un link a questo documento: https://hdl.handle.net/2108/404166
Citazioni
  • ???jsp.display-item.citation.pmc??? ND
  • Scopus 0
  • ???jsp.display-item.citation.isi??? ND
social impact