Visual decoding of neurophysiological signals is pivotal for Brain-Computer Interfaces (BCIs) and AI. Current approaches often fail to bridge the systematic gap and stochastic gap between neural and visual modalities, neglecting the intrinsic mechanisms of the Human Visual System (HVS). To bridge these gap, we propose Brain-Inspired Capture (BI-Cap), a \textbf{Neuromimetic Perceptual Simulation} paradigm that aligns modalities based on HVS processing. We implement four dynamic and static biologically plausible transformations, explicitly incorporating MI-guided dynamic blur regulation to simulate adaptive visual processing. Furthermore, addressing the inherent high dynamicity and inter-subject heterogeneity of neural activity, we introduce an \textbf{Evidence-Driven Latent Space Representation} framework. This approach facilitates robust neural representation by explicitly modeling the uncertainty within HVS processing. Extensive experiments on zero-shot brain-to-image retrieval tasks across two public benchmark datasets demonstrate the effectiveness of our paradigm, surpassing state-of-the-art methods by significant margins of 9.2/% and 8.0/%, respectively.
- OS: Linux
- CUDA: 11.8
- Python: 3.11.13
- Pytorch: 2.4.1
pip install -r requirements.txt./data
├── things_eeg #Preprocessed THINGS-EEG2 dataset for train and test
│ └── Preprocessed_data_250Hz_whiten
│ ├── sub-01
│ │ ├── train.npy
│ │ └── test.npy
│ ├── sub-02
│ │ ├── train.npy
│ │ └── test.npy
│ └── ...
├── things_meg #Preprocessed THINGS-MEG dataset for train and test
│ └── Preprocessed_data
│ ├── sub-01
│ │ ├── train.npy
│ │ └── test.npy
│ ├── sub-02
│ │ ├── train.npy
│ │ └── test.npy
│ └── ...
└── images_set #images dataset for train and test
│ └── test_images
│ └── training_images
└── images_set_resize #Preprocessed images dataset for train and test
│ └── test_images
│ └── training_images
└── Image_set_Gaussian #Image transformation dataset for train and test
│ └── test_images
│ └── training_images
└── Image_set_LowRes #Image transformation dataset for train and test
│ └── test_images
│ └── training_images
└── Image_set_Mosaic #Image transformation dataset for train and test
│ └── test_images
│ └── training_images
python ./data_preprocessing/process_eeg_whiten.py
python ./data_preprocessing/process_meg.pypython ./data_preprocessing/process_resize.py
python ./data_preprocessing/Static_transformation.pypython main-eeg.py --config configs/BI-Cap_eeg.yaml --subjects sub-01 --seed 0 --exp_setting intra-subject --brain_backbone Shared_Temporal_Attention_Encoder --vision_backbone RN50 --adaptor_backbone ShrinkAdapter --epoch 150 --lr 1e-4 --device cuda:0 --train_batch_size 32We acknowledge the contributions of the following datasets:
- A large and rich EEG dataset for modeling human visual object recognition) [THINGS-EEG]
- THINGS-data, a multimodal collection of large-scale datasets for investigating object representations in human brain and behavior](https://pubmed.ncbi.nlm.nih.gov/36847339) [THINGS-MEG] The inspiration for this code comes from previous excellent work:
- [Bridging the Vision-Brain Gap with an Uncertainty-Aware Blur Prior](https://openreview.net/forum?id=38qCgXJ8L0](https://github.com/HaitaoWuTJU/Uncertainty-aware-Blur-Prior) [CVPR 2025]
- Shrinking the Teacher: An Adaptive Teaching Paradigm for Asymmetric EEG-Vision Alignment [AAAI 2026]
