MindTuner:
Cross-Subject Visual Decoding with Visual Fingerprint and Semantic Correction
Abstract
Decoding natural visual scenes from brain activity has flourished, with extensive research in single-subject tasks and, however, less in cross-subject tasks. Reconstructing high-quality images in cross-subject tasks is a challenging problem due to profound individual differences between subjects and the scarcity of data annotation. In this work, we proposed MindTuner for cross-subject visual decoding, which achieves high-quality and rich semantic reconstructions using only 1 hour of fMRI training data benefiting from the phenomena of visual fingerprint in the human visual system and a novel fMRI-to-text alignment paradigm. Firstly, we pre-train a multi-subject model among 7 subjects and fine-tune it with scarce data on new subjects, where LoRAs with Skip-LoRAs are utilized to learn the visual fingerprint. Then, we take the image modality as the intermediate pivot modality to achieve fMRI-to-text alignment, which achieves impressive fMRI-to-text retrieval performance and corrects fMRI-to-image reconstruction with fine-tuned semantics. The results of both qualitative and quantitative analyses demonstrate that MindTuner surpasses state-of-the-art cross-subject visual decoding models on the Natural Scenes Dataset (NSD), whether using training data of 1 hour or 40 hours.
Introduction
Do our brains form unified perceptions when we observe similar objects? Do our unique understandings influence these perceptions differently? The human brain exhibits substantial anatomical similarities in terms of functional organization, including shared attributes like memory, functional connectivity, and visual cortex functions (Chen et al. 2017; Fingelkurts, Fingelkurts, and Kähkönen 2005; Stringer et al. 2019). However, individual neural biases always exist due to inherent differences (Wang, Murai, and Whitney 2020). Understanding both the similarities and gaps in perception has profound implications for the fields of Artificial Intelligence (AI) (Nie et al. 2023; Zhao et al. 2014) and Brain-Computer Interface (BCI) research (Wang et al. 2009). Visual decoding is a straightforward way to understand the brain, where functional magnetic resonance imaging (fMRI) is a widely embraced non-invasive tool used to decode natural visual stimuli, revealing intricate perceptual and semantic details in the cerebral cortex (Linden 2021). Consequently, fMRI has garnered considerable attention in image retrieval and reconstruction tasks.
Open-source large-scale fMRI datasets, such as the Natural Scenes Dataset (NSD) (Allen et al. 2022), advance deep learning models to shine in fMRI decoding. Pre-trained cross-modality models like CLIP (Radford et al. 2021) and Stable Diffusion (Rombach et al. 2022) offer effective representation space and models for high-quality visual reconstruction. A large body of literature demonstrates the feasibility of training single-subject decoding models to reconstruct high-fidelity images (Lin, Sprague, and Singh 2022; Scotti et al. 2023; Chen et al. 2023; Takagi and Nishimoto 2023; Lu et al. 2023). However, single-subject decoding has drawbacks, including the need to train a unique model for each subject, making it challenging to generalize to new subjects and requiring a substantial amount of fMRI data for training. As is widely recognized, acquiring a large amount of fMRI data for each subject is time-consuming, labor-intensive, and impractical in practical scenarios. Unfortunately, most current research focuses on single-subject visual decoding rather than exploring the challenging commonalities of the human brain. Consequently, there is an urgent need for cross-subject decoding models that can be effectively transferred to new subjects and perform well in the few-shot setting, as depicted in Figure LABEL:1_hour.
The key to cross-subject few-shot decoding lies in effectively utilizing extensive prior knowledge from other subjects or additional modalities. On the one hand, one successful strategy for leveraging knowledge from other subjects involves aligning them to a shared space. Ridge regression is commonly employed for this purpose, aligning voxel inputs from different subjects (Scotti et al. 2024; Ferrante, Boccato, and Toschi 2023). This approach is preferred due to the low signal-to-noise ratio in fMRI data, where complex non-linear models tend to overfit noise. Nonetheless, the process of visual information perception and generating brain activity in each individual incorporates unique components, referred to as the visual fingerprint (Wang, Murai, and Whitney 2020) (refer to Preliminary for more detailed analysis). Current linear alignment methods only enable new subjects to conform to the shared components across subjects, neglecting the perception difference derived from their distinctive visual fingerprint and resulting in limited performance.
On the other hand, an additional strategy involves leveraging multimodal data. Previous alignment methods focused solely on the visual modality, as a means of adapting to the inputs of Stable Diffusion. However, this alignment approach is susceptible to slight disturbances, leading to semantic errors in the generated images. For visual decoding tasks, the textual modality is highly relevant and is verified effective in enhancing visual decoding semantically (Scotti et al. 2024). However, previous approaches incorporating text have placed excessive emphasis on directly aligning fMRI with detailed textual descriptions to facilitate the reconstruction process (Takagi and Nishimoto 2023), which intuitively lacks rationality and yields poor performance. Considering that subjects in visual stimulation experiments lack direct interaction with the textual modality and that individual understanding of visual stimuli varies, the relationship between fMRI and text should be considered implicit.
In this paper, we propose MindTuner: a cross-subject visual decoding framework. Inspired by visual fingerprint in Brain Science (Wang, Murai, and Whitney 2020), and with the help of Low-Rank Adaptation (LoRA) for large model lightweight fine-tuning. In correspondence to the above two strategies, we first propose the combination of non-linear Skip-LoRAs and LoRAs to learn the visual fingerprint of new subjects, which are injected into the fMRI encoding network to correct visual perception difference. In addition, we design a Pivot module that uses images as the central modality to bridge fMRI and text. The Pivot helps to correct the reconstructed image with fine-tuned semantics. Our contributions are summarized as follows:
-
•
MindTuner makes the first attempt to introduce LoRA as a subject-level fine-tuning module in cross-subject decoding and further elaborately design non-linear Skip-LoRA. Their combination shows excellent capability to learn subjects’ visual fingerprint.
-
•
We introduce a novel fMRI-to-text retrieval paradigm with a Pivot using the image modality. The Pivot conducts semantic correction with label prompts to enhance fMRI-to-image reconstruction.
-
•
We evaluate our method on the NSD dataset, and it establishes SOTA decoding performance whether using training data of 40 hours (full) or 1 hour(2.5% of full).
Related Work
Cross-Subject Functional Alignment
In visual decoding, single-subject models have exposed the issue of excessive reliance on the data volume of individual subjects, leading researchers to shift toward cross-subject studies. Functional alignment of different brains is considered to be more effective than anatomical alignment. Previous functional alignment methods were mainly divided into two perspectives: fMRI data itself and downstream tasks. Among a series of methods starting from the fMRI perspective, Bazeille et al. (Bazeille et al. 2019) and Thual et al. (Thual et al. 2023) minimize an optimal transport cost between voxels of different brains. The methods based on fMRI self-supervision (Chen et al. 2023; Qian et al. 2023) emphasize obtaining common latent representations between subjects through autoencoders. Wills Aligner (Bao et al. 2024) aligns data from different subjects using the fsaverage template. On another cross-subject alignment research path, most of these methods focus on the shared knowledge among subjects and overlook the complex nonlinear relationships between subjects due to concerns about overfitting. Linear fitting (Ferrante, Boccato, and Toschi 2023) was performed on the responses of subjects to common images, achieving high-quality cross-subject visual reconstruction, but requiring subjects to see the same images. Mindeye2 (Scotti et al. 2024), MindBridge (Wang et al. 2024) and UMBRAE (Xia et al. 2024b) improved this and achieved impressive results, but still losing subject-specific features and ignoring the more complex relationships between subjects, which is not flexible.
Text Modality in Visual Decoding
Decoding visual stimuli from fMRI has been a long-standing endeavor, primarily focusing on the image modality (Takagi and Nishimoto 2023; Lu et al. 2023; Gong et al. 2024b; Ozcelik and VanRullen 2023; Scotti et al. 2023). However, an increasing number of studies have highlighted the role of text. UniBrain (Mai and Zhang 2023) directly aligns fMRI to the text representation via ridge regression and then completes the brain caption task with the text generation model. MindEye2 (Scotti et al. 2024) obtains a better caption by placing the generated image representation into the image captioning model, which is used to smooth the generated image. NeuroClips utilizes the pretrained BLIP-2 (Li et al. 2023) model to achieve text-assisted video reconstruction (Gong et al. 2024a). However, these methods either do direct fMRI-to-text alignment or direct image-text alignment, ignoring the indirect relationship between fMRI and text from the perspective of intuitive understanding.
Preliminary on visual fingerprint
Visual stimuli are processed by the Human Visual System(HVS). Therefore, as stated in previous research (Xia et al. 2024a), brain responses such as fMRI are closely linked to our visual system. Although presented in a systematic manner, there are still significant differences in the visual systems of different individuals. Research of visual fingerprint emphasizes that idiosyncratic biases exist in the underlying representation of visual space propagate across varying levels of visual processing (Wang, Murai, and Whitney 2020). Using a position-matching task will find stable subject-specific compressions and expansions within local regions throughout the visual field. As shown on the left of Figure 1, in experiments, subjects were fixated at the center, and a target was displayed briefly at one of five possible eccentricities (depicted by dashed lines, which were not visible in the experiment). After the target disappeared, subjects moved the cursor to match the target’s location. Linear models were used to fit the Distortion Indices(DI), which measured the degree of spatial distortion between subjects:
(1) | ||||
As shown on the right of Figure 1, the experiment results show that subjects have their own visual fingerprint, with both linear and non-linear components. Within-subject similarity (r=0.71) is significantly higher than between-subject similarity (r=0.22), suggesting that each individual subject has their own unique spatial distortions that are consistent within themselves and distinguished from others.

Method

The task of cross-subject visual decoding consists of two parts: multi-subject pre-training and new-subject fine-tuning. Given a neural dataset with brain activities of subject , it involves the reconstruction of visual images through a core pipeline whose inputs are flattened and aligned fMRI voxels after ROI extraction, where denotes voxel numbers. To simplify notation, each denotes data from the original subjects for this section. Our goal is to optimize the , so that , where best approximates . New-subject fine-tuning adheres to the same path, albeit with scarce data. For the multi-subject pre-training, we follow the pipeline of MindEye2, while for the new-subject fine-tuning, we plug into visual fingerprint and Pivot as shown in Figure 2.
Multi-Subject Pre-training
Multi-Subject Functional Alignment
It should be noted that the brain structures of different subjects vary, as do the number of voxels obtained. Consequently, a mapping model is required to align the voxel inputs from different subjects to the same dimension as the inputs to the model in the phase of multi-subject pre-training. Here, we employ linear function alignment to learn shared-subject fMRI latent space ( denotes shared input dimension). This is achieved through subject-specific ridge regression, as detailed below:
(2) |
MLP Backbone
In order to achieve high-fidelity reconstruction, it is necessary to utilize a substantial quantity of CLIP image embedding. The OpenCLIP ViT-bigG/14 space is employed for alignment, with an image embedding dimension of . Mapped inputs are fed into an MLP backbone comprising four residual blocks and a tokenization layer which transforms the input from a dimension of to . The MLP Backbone serves to convert the fMRI to the intermediate backbone embedding space: . It should be noted that all subjects shared the same MLP backbone following multi-subject functional alignment. Subsequently, the backbone embeddings are conveyed into three submodules for retrieval, high-level reconstruction, and low-level reconstruction.
Retrieval Submodules
A straightforward approach to performing the retrieval task is to conduct a shallow mapping of the backbone embeddings and supervise it with an fMRI-to-image CLIP contrastive loss. In the case of limited fMRI data, the application of appropriate data augmentation techniques can facilitate the convergence of the model. A recently proposed voxel mixture paradigm, based on MixCo (Kim et al. 2020), has demonstrated effectiveness. Two raw fMRI voxels and are mixed into using a factor sampled from the Beta distribution:
(3) | ||||
where denotes an arbitrary mixing index in the batch. The forward mixed contrastive loss MixCo is formulated as:
(4) | |||
where denotes ground-truth CLIP image embeddings, denotes a temperature hyperparameter, and is the batch size. Here we use the bidirectional loss .
Low-level and High-level Submodules
The low-level pipeline is a widely utilized technique for the enhancement of low-level visual metrics in reconstruction images. This involves the mapping of voxels to the latent space of Stable Diffusion’s Variational AutoEncoder (SDVAE), which serves as a surrogate for the reconstruction. The pipeline comprises an MLP and a CNN upsampler with L1 loss in Stable Diffusion’s latent embeddings .
(5) |
Conversely, a high-level pipeline places greater emphasis on semantic alignment. Inspired by DALLE·2 (Ramesh et al. 2022), a diffusion prior is recognized as an effective means of transforming backbone embeddings into CLIP ViT image embeddings, in which mean square error loss is used (further insights on the deployment of diffusion priors can be found in Appendix B.4):
(6) |
Thus, the end-to-end loss for multi-subject pre-training is:
(7) |
Where denotes the weight to balance multiple losses.
New-Subject Fine-tuning
Low-Rank Adaptation
Previous work has demonstrated the effectiveness of low-rank adaptation in fine-tuning large language models with significantly fewer trainable parameters. It is well suited to our multi-subject decoding task for two reasons. First, most of the current mainstream models for fMRI decoding are MLP-based models that contain a large number of linear layers, whereas LoRA has been shown to achieve good fine-tuning results in the linear layers. Second, in cross-subject scenarios, fMRI data from new subjects are usually scarce, and full-tuning the whole model is usually difficult to grasp, leading to a certain degree of overfitting. For each pre-trained weight matrix in the multi-subject model, where denotes input dimension and denotes output dimension, gradient update is constrained with a low-rank decomposition for new-subject adapter matrix :
(8) |
where is keeped frozen, , , is the rank and . At the beginning of the training phase, the parameters of the matrix are randomly initialized, and is initialized to zero, which ensures that the initial output of the LoRA block is all zeros.
non-linear Skip-LoRAs
The LoRA model is notably lightweight, effectively circumvents complexity, and could avoid the aforementioned fMRI overfitting problem. Although LoRA can be an effective means of fine-tuning multi-subject models, simple linear LoRA models are insufficient for the capture of visual fingerprints. Indeed, there exists a non-linear relationship between subjects (discussed in the Preliminary Section). Therefore, we inject a non-linear design into it. Here, we design our non-linear LoRAs by adding the activation function and the nonlinear constraints. Inspired by the skip-connection of Unet in Computer Vision (Zhang, Rao, and Agrawala 2023), we built a brand new Skip-LoRAs to bootstrap the initial non-linearity of fMRI between subjects directly affecting the entire MLP backbone(see Appendix B.2 for Skip-LoRAs). Skip-LoRAs can be defined as . Assuming that is the -th layer of MLP backbone , the output of the new-subject backbone could be as follows:
(9) | ||||
Here, we use the Pearson correlation coefficient to define the non-linear correlation loss:
(10) |
Pivot with Adaptive Projector
In addition to pixel-level alignment, semantic-level matching is equally important for reconstructing the semantic integrity of an image. In visual decoding tasks, fMRI is directly associated with the image stimulus, and semantic information is implicit in fMRI. More details on the Adaptive Projector can be found in Appendix B.3. For new subjects, we add the additional image-text loss to the model to constrain the semantic information:
(11) |
where denotes the projector’s output from image tokens , and denotes the CLS embeddings (CLS denotes the classification token) of paired text. During new-subject fine-tuning, the adaptive projector is trainable. Thus, the end-to-end loss for new-subject fine-tuning is:
(12) |
Semantic Correction
MindEye2 has found that the reconstructed images from SDXL unCLIP after token alignment have fuzzy semantics, so the aligned embeddings are fed into the image captioning model GIT (Wang et al. 2022) to get the semantics for refinement. We have found that single category words are sufficient to accomplish the refinement task. In addition, the captions generated by MindEye2 are completely dependent on aligned embeddings, which can only make the image semantics clearer and cannot correct the semantics of the reconstructed images, as shown in Figure 3. Taking advantage of our adaptive projector, we can define category description as an fMRI-to-text retrieval task using a simple text prompt . In this way, we achieve more accurate category reconstruction.

Method | Low-Level | High-Level | Retrieval | |||||||
---|---|---|---|---|---|---|---|---|---|---|
PixCorr | SSIM | Alex(2) | Alex(5) | Incep | CLIP | Eff | SwAV | Image | Brain | |
Takagi… | 0.246 | 0.410 | 78.9% | 85.6% | 83.8% | 82.1% | 0.811 | 0.504 | - | - |
Ozcelik… | 0.273 | 0.365 | 94.4% | 96.6% | 91.3% | 90.9% | 0.728 | 0.422 | 18.8% | 26.3% |
MindEye1 | 0.319 | 0.360 | 92.8% | 96.9% | 94.6% | 93.3% | 0.648 | 0.377 | 90.0% | 84.1% |
MindEye2 | 0.322 | 0.431 | 96.1% | 98.6% | 95.4% | 93.0% | 0.619 | 0.344 | 98.8% | 98.3% |
MindTuner(Ours) | 0.322 | 0.421 | 95.8% | 98.8% | 95.6% | 93.8% | 0.612 | 0.340 | 98.9% | 98.3% |
MindEye2(1 hour) | 0.195 | 0.419 | 84.2% | 90.6% | 81.2% | 79.2% | 0.810 | 0.468 | 79.0% | 57.4% |
MindTuner(1 hour) | 0.224 | 0.420 | 87.8% | 93.6% | 84.8% | 83.5% | 0.780 | 0.440 | 83.1% | 76.0% |
Experiment
Datasets
Natural Scenes Dataset (NSD)111https://naturalscenesdataset.org (Allen et al. 2022) is an extensive 7T fMRI dataset gathered from 8 subjects viewing images from the MSCOCO-2017 dataset, which contains images of complex natural scenes. Participants viewed three repetitions of 10,000 images with a 7-Tesla fMRI scanner over 30–40 sessions, with one session including 750 fMRI trials lasting for 1 hour. Details of the dataset information can be found in Appendix A. In this paper, we conducted 4 types of experiments, image retrieval and reconstruction in the next section, and text retrieval and brain correlation experiments in Appendix C.
Implementation details
All of our fine-tuning experiments were run for 150 epochs on two Tesla v100 32GB GPUs with a batch size of 10. The experimental parameter settings for 1-hour and 40-hour data are consistent. For fine-tuning experiments, the loss weight is set to and the rank is set to 8 for all LoRA blocks, including Skip-LoRAs. We use the AdamW (Loshchilov and Hutter 2017) for optimization, with a learning rate set to 3e-4, to which the OneCircle learning rate schedule (Smith and Topin 2019) was set. During the training process, we use data augmentation from images and blurry images and replace the BiMixCo with SoftCLIP (Scotti et al. 2023) loss in one-third of the training phase. In the inference stage, we only generated a reconstructed image once and did not make multiple selections. The final high-level reconstructed image and low-level blurry image are simply weighted and averaged in a ratio of 3:1. In the inference stage of semantic correction, we use 80 category nouns from MSCOCO as correction texts and classify them in the form of text retrieval.
Results and Analysis
Image and Brain Retrieval
Image retrieval refers to retrieving the image embeddings with the highest cosine similarity based on fMRI embeddings on the test set. In Table 1, there is only a slight improvement in retrieval accuracy within 40 hours of data. This can be attributed to the fact that the retrieval accuracy is already close to the upper limit (only about 1% short of reaching 100%), which is affected by the noise of the fMRI dataset itself. However, when only 1 hour of data was available, MindTuner’s retrieval accuracy was significantly higher than MineEye2, 4.1% higher for image retrieval and 18.6% higher for brain retrieval. As MindEye2 benefits from multi-subject pre-training and new-subject fine-tuning with linear heads, this suggests that the visual fingerprint we introduced significantly improves the performance of the model when fMRI data are scarce.


Method | Trainable | Low-Level | High-Level | Retrieval | |||||||
---|---|---|---|---|---|---|---|---|---|---|---|
Parameters | PixCorr | SSIM | Alex(2) | Alex(5) | Incep | CLIP | Eff | SwAV | Image | Brain | |
MindEye2 | 64.4M(-) | 0.235 | 0.428 | 88.0% | 93.3% | 83.6% | 80.8% | 0.798 | 0.459 | 94.0% | 77.6% |
with Adaptive Projector | 66.5M(+2.1M) | 0.233 | 0.426 | 87.8% | 93.0% | 84.0% | 81.2% | 0.794 | 0.454 | 93.8% | 77.3% |
with only LoRAs | 74.6M(+9.4M) | 0.261 | 0.427 | 90.3% | 94.2% | 84.8% | 84.0% | 0.784 | 0.441 | 93.7% | 85.6% |
LoRAs+Skip-LoRAs | 74.6M(+10.2M) | 0.264 | 0.427 | 90.8% | 94.8% | 85.1% | 84.2% | 0.780 | 0.437 | 94.5% | 87.7% |
MindTuner(non-linear) | 76.7M(+12.3M) | 0.191 | 0.383 | 82.7% | 88.0% | 77.3% | 76.0% | 0.848 | 0.492 | 27.4% | 56.7% |
MindTuner | 76.7M(+12.3M) | 0.262 | 0.422 | 90.6% | 94.9% | 85.8% | 84.6% | 0.774 | 0.433 | 94.2% | 87.4% |
Trainable | Skip-LoRAs | Low-Level | High-Level | Retrieval | ||||||||
---|---|---|---|---|---|---|---|---|---|---|---|---|
Parameters | Parameters | PixCorr | SSIM | Alex(2) | Alex(5) | Incep | CLIP | Eff | SwAV | Image | Brain | |
71.6M(+7.2M) | 0.4M | 0.263 | 0.422 | 90.4% | 94.9% | 84.8% | 84.6% | 0.778 | 0.436 | 94.4% | 87.9% | |
76.7M(+12.3M) | 0.8M | 0.262 | 0.422 | 90.6% | 94.9% | 85.8% | 84.6% | 0.774 | 0.433 | 94.2% | 87.4% | |
86.7M(+22.3M) | 1.6M | 0.262 | 0.422 | 90.5% | 95.1% | 85.2% | 85.2% | 0.776 | 0.434 | 93.7% | 87.2% |
Image Reconstruction
Image reconstruction aims to restore the original image as seen by the subjects, and two levels of evaluation metrics assess the quality of the reconstructions. Previous research has improved the performance of these above metrics in single-subject models by various means. The cross-subject task discussed in this paper tests the ability of the model to exploit the commonalities of multi-subject and to migrate new subjects, ultimately realizing the goal of using less fMRI data for few-shot learning. Here, we report the reconstruction results of MindTuner using 1-hour and 40-hour training data. Table 1 reflects the quantized performance comparisons, under two different training data sizes. It can be seen that when using the full NSD dataset, MindTuner achieves better performance on high-level metrics; when only 1 hour of data is available for training, MindTuner outperforms MindEye2 on all metrics. The visualization results of 1-hour and 40-hour reconstructions can be seen in Figure 4 and Figure 5. It can be seen that the quality of the reconstructed images at 40 hours is significantly higher than at 1 hour as the training data increases. Meanwhile, our MindTuner is better than MindEye2 in both semantic completeness and category accuracy, demonstrating the superiority of the overall model. Our visualization excludes the other three methods in Table 1 because the image semantics generated by these three methods are very unclear. Interestingly, we also found that images generated directly from SDXL unCLIP outperformed MindTuner’s corrected images at a high level, albeit visibly distorted. Further results are in Appendix D.1.
Ablations
In this section, we explored the effectiveness of each component of our method through ablation experiments. All results were pre-trained on Subjects 2-8, and fine-tuning was performed on Subject 1 using 1 hour of data.
MindTuner’s modules. Further experiments were conducted to assess the efficacy of each module. As can be observed in Table 2, the incorporation of LoRAs markedly enhances the model’s capacity. The incorporation of Skip-LoRAs resulted in a convergence of all image reconstruction metrics towards the performance of the complete model, with the retrieval accuracy even exceeding that of the complete model. This can be attributed to the balance of additional image-to-text aligned losses, which has the effect of slightly reducing the performance of the retrieval submodule in the complete model. Furthermore, the incorporation of an adaptive projector has enhanced the performance of high-level reconstruction, and additional visualizations in Figure 3 have also demonstrated the efficacy of semantic correction. However, this approach resulted in the compromise of certain low-level effects. In comparison to the 2.2B multi-subject pre-trained main model, MindTuner resulted in a mere 12.3M increase in parameters (0.56%), yet achieved superior outcomes. Furthermore, an activation function was added to the ridge head to make the entire alignment head non-linear. However, this resulted in a serious overfitting phenomenon, with all metrics exhibiting a significant decline, which indicates that the Skip-LoRAs in MindTuner offer a more suitable alternative for non-linear fMRI alignment of diverse subjects.
The rank of Skip-LoRAs and LoRAs. Further experiments were conducted on the most critical variable in the LoRAs, namely the rank . As can be observed in Table 3, when the rank , the overall ability of the model does not show significant changes. This finding is consistent with the results presented in the LoRA paper (Hu et al. 2021), which highlighted that when r is within the range of 2 to 16, the fine-tuning effect remains unchanged. Furthermore, no substantial evidence of overfitting was identified. This can be attributed to the fact that the parameters in Skip-LoRAs have remained relatively modest, below 2M.
Neuroscience interpretability
To investigate the interpretability of MindTuner in neuroscience, we conduct experiments to explore where the subject’s non-linear relationship comes from in the visual cortex. We utilize pycortex (Gao et al. 2015) to project the weights of each voxel in the first layer of Ridge regression, LoRA, or Skip-LoRA onto the corresponding 2D flat map of the NSD dataset. The results are presented in Figure 6. For linear heads, the importance of visual cortical voxels is similar for both ridge and LoRA. However, for Skip-LoRA, a small portion of the early visual cortex and more advanced visual cortex are valued more. This indicates that at the fMRI level, a greater concentration of non-linear relationships is found within the higher visual cortex. It also demonstrates that the design of Skip-LoRAs is capable of capturing non-linear relationships in fMRI data that are not discernible with standard LoRAs. More visualization results, as well as brain region ROI templates, can be found in Appendix E.

Conclusion
In this paper, we propose MindTuner, a new cross-subject decoding method. We introduced the phenomenon of visual fingerprint in the human visual system and utilized the combination of Skip-LoRAs and LoRAs to learn each subject’s visual fingerprint. Meanwhile, we innovatively propose a method for enhancing reconstruction by indirectly connecting fMRI with text in visual decoding tasks. Experimental results have shown that we have achieved better performance on multiple evaluation metrics at a relatively small parameter cost, especially when the fMRI data is insufficient. Our work has relaxed the conditions for fMRI acquisition, helping to achieve a universal brain decoding model in the future.
Acknowledgements
We have included more experimental results and analyses in the appendix, which you can access through this arXiv link222https://arxiv.org/abs/2404.12630.
This research is supported by the National Key Research and Development Program of China (No. 2022YFB3104700), the National Natural Science Foundation of China (No. 61976158, No. 62376198), Shanghai Baiyulan Pujiang Project (No. 08002360429).
References
- Allen et al. (2022) Allen, E. J.; St-Yves, G.; Wu, Y.; Breedlove, J. L.; Prince, J. S.; Dowdle, L. T.; Nau, M.; Caron, B.; Pestilli, F.; Charest, I.; et al. 2022. A massive 7T fMRI dataset to bridge cognitive neuroscience and artificial intelligence. Nature neuroscience, 25(1): 116–126.
- Bao et al. (2024) Bao, G.; Gong, Z.; Zhang, Q.; Zhou, J.; Fan, W.; Yi, K.; Naseem, U.; Hu, L.; and Miao, D. 2024. Wills Aligner: A Robust Multi-Subject Brain Representation Learner. arXiv preprint arXiv:2404.13282.
- Bazeille et al. (2019) Bazeille, T.; Richard, H.; Janati, H.; and Thirion, B. 2019. Local optimal transport for functional brain template estimation. In Information Processing in Medical Imaging: 26th International Conference, IPMI 2019, Hong Kong, China, June 2–7, 2019, Proceedings 26, 237–248. Springer.
- Caron et al. (2020) Caron, M.; Misra, I.; Mairal, J.; Goyal, P.; Bojanowski, P.; and Joulin, A. 2020. Unsupervised learning of visual features by contrasting cluster assignments. Advances in neural information processing systems, 33: 9912–9924.
- Chen et al. (2017) Chen, J.; Leong, Y. C.; Honey, C. J.; Yong, C. H.; Norman, K. A.; and Hasson, U. 2017. Shared memories reveal shared structure in neural activity across individuals. Nature neuroscience, 20(1): 115–125.
- Chen et al. (2023) Chen, Z.; Qing, J.; Xiang, T.; Yue, W. L.; and Zhou, J. H. 2023. Seeing beyond the brain: Conditional diffusion model with sparse masked modeling for vision decoding. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 22710–22720.
- Ferrante, Boccato, and Toschi (2023) Ferrante, M.; Boccato, T.; and Toschi, N. 2023. Through their eyes: multi-subject Brain Decoding with simple alignment techniques. arXiv preprint arXiv:2309.00627.
- Fingelkurts, Fingelkurts, and Kähkönen (2005) Fingelkurts, A. A.; Fingelkurts, A. A.; and Kähkönen, S. 2005. Functional connectivity in the brain—is it an elusive concept? Neuroscience & Biobehavioral Reviews, 28(8): 827–836.
- Gao et al. (2015) Gao, J. S.; Huth, A. G.; Lescroart, M. D.; and Gallant, J. L. 2015. Pycortex: an interactive surface visualizer for fMRI. Frontiers in neuroinformatics, 9: 23.
- Gong et al. (2024a) Gong, Z.; Bao, G.; Zhang, Q.; Wan, Z.; Miao, D.; Wang, S.; Zhu, L.; Wang, C.; Xu, R.; Hu, L.; et al. 2024a. NeuroClips: Towards High-fidelity and Smooth fMRI-to-Video Reconstruction. arXiv preprint arXiv:2410.19452.
- Gong et al. (2024b) Gong, Z.; Zhang, Q.; Bao, G.; Zhu, L.; Zhang, Y.; Liu, K.; Hu, L.; and Miao, D. 2024b. Lite-Mind: Towards Efficient and Robust Brain Representation Learning. In Proceedings of the 32nd ACM International Conference on Multimedia, MM ’24, 4014–4023. New York, NY, USA: Association for Computing Machinery. ISBN 9798400706868.
- Hu et al. (2021) Hu, E. J.; Shen, Y.; Wallis, P.; Allen-Zhu, Z.; Li, Y.; Wang, S.; Wang, L.; and Chen, W. 2021. Lora: Low-rank adaptation of large language models. arXiv preprint arXiv:2106.09685.
- Kim et al. (2020) Kim, S.; Lee, G.; Bae, S.; and Yun, S.-Y. 2020. Mixco: Mix-up contrastive learning for visual representation. arXiv preprint arXiv:2010.06300.
- Li et al. (2023) Li, J.; Li, D.; Savarese, S.; and Hoi, S. 2023. Blip-2: Bootstrapping language-image pre-training with frozen image encoders and large language models. In International conference on machine learning, 19730–19742. PMLR.
- Lin, Sprague, and Singh (2022) Lin, S.; Sprague, T.; and Singh, A. K. 2022. Mind reader: Reconstructing complex images from brain activities. Advances in Neural Information Processing Systems, 35: 29624–29636.
- Linden (2021) Linden, D. 2021. Section 3 - Introduction. In Hampson, M., ed., fMRI Neurofeedback, 161–169. Academic Press. ISBN 978-0-12-822421-2.
- Loshchilov and Hutter (2017) Loshchilov, I.; and Hutter, F. 2017. Decoupled weight decay regularization. arXiv preprint arXiv:1711.05101.
- Lu et al. (2023) Lu, Y.; Du, C.; Zhou, Q.; Wang, D.; and He, H. 2023. MindDiffuser: Controlled Image Reconstruction from Human Brain Activity with Semantic and Structural Diffusion. In Proceedings of the 31st ACM International Conference on Multimedia, 5899–5908.
- Mai and Zhang (2023) Mai, W.; and Zhang, Z. 2023. Unibrain: Unify image reconstruction and captioning all in one diffusion model from human brain activity. arXiv preprint arXiv:2308.07428.
- Nie et al. (2023) Nie, X.; Hu, B.; Gao, X.; Li, L.; Zhang, X.; and Xiao, B. 2023. BMI-Net: A Brain-inspired Multimodal Interaction Network for Image Aesthetic Assessment. In Proceedings of the 31st ACM International Conference on Multimedia, MM ’23, 5514–5522. New York, NY, USA: Association for Computing Machinery. ISBN 9798400701085.
- Ozcelik and VanRullen (2023) Ozcelik, F.; and VanRullen, R. 2023. Brain-diffuser: Natural scene reconstruction from fmri signals using generative latent diffusion. arXiv preprint arXiv:2303.05334.
- Qian et al. (2023) Qian, X.; Wang, Y.; Huo, J.; Feng, J.; and Fu, Y. 2023. fmri-pte: A large-scale fmri pretrained transformer encoder for multi-subject brain activity decoding. arXiv preprint arXiv:2311.00342.
- Radford et al. (2021) Radford, A.; Kim, J. W.; Hallacy, C.; Ramesh, A.; Goh, G.; Agarwal, S.; Sastry, G.; Askell, A.; Mishkin, P.; Clark, J.; et al. 2021. Learning transferable visual models from natural language supervision. In International conference on machine learning, 8748–8763. PMLR.
- Ramesh et al. (2022) Ramesh, A.; Dhariwal, P.; Nichol, A.; Chu, C.; and Chen, M. 2022. Hierarchical text-conditional image generation with clip latents. arXiv preprint arXiv:2204.06125.
- Rombach et al. (2022) Rombach, R.; Blattmann, A.; Lorenz, D.; Esser, P.; and Ommer, B. 2022. High-resolution image synthesis with latent diffusion models. In Proceedings of the IEEE/CVF conference on computer vision and pattern recognition, 10684–10695.
- Scotti et al. (2023) Scotti, P. S.; Banerjee, A.; Goode, J.; Shabalin, S.; Nguyen, A.; Cohen, E.; Dempster, A. J.; Verlinde, N.; Yundler, E.; Weisberg, D.; et al. 2023. Reconstructing the Mind’s Eye: fMRI-to-Image with Contrastive Learning and Diffusion Priors. arXiv preprint arXiv:2305.18274.
- Scotti et al. (2024) Scotti, P. S.; Tripathy, M.; Villanueva, C. K. T.; Kneeland, R.; Chen, T.; Narang, A.; Santhirasegaran, C.; Xu, J.; Naselaris, T.; Norman, K. A.; et al. 2024. MindEye2: Shared-Subject Models Enable fMRI-To-Image With 1 Hour of Data. arXiv preprint arXiv:2403.11207.
- Smith and Topin (2019) Smith, L. N.; and Topin, N. 2019. Super-convergence: Very fast training of neural networks using large learning rates. In Artificial intelligence and machine learning for multi-domain operations applications, volume 11006, 369–386. SPIE.
- Stringer et al. (2019) Stringer, C.; Pachitariu, M.; Steinmetz, N.; Carandini, M.; and Harris, K. D. 2019. High-dimensional geometry of population responses in visual cortex. Nature, 571(7765): 361–365.
- Takagi and Nishimoto (2023) Takagi, Y.; and Nishimoto, S. 2023. High-resolution image reconstruction with latent diffusion models from human brain activity. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 14453–14463.
- Tan and Le (2019) Tan, M.; and Le, Q. 2019. Efficientnet: Rethinking model scaling for convolutional neural networks. In International conference on machine learning, 6105–6114. PMLR.
- Thual et al. (2023) Thual, A.; Benchetrit, Y.; Geilert, F.; Rapin, J.; Makarov, I.; Banville, H.; and King, J.-R. 2023. Aligning brain functions boosts the decoding of visual semantics in novel subjects. arXiv preprint arXiv:2312.06467.
- Wang et al. (2009) Wang, J.; Pohlmeyer, E.; Hanna, B.; Jiang, Y.-G.; Sajda, P.; and Chang, S.-F. 2009. Brain state decoding for rapid image retrieval. In Proceedings of the 17th ACM International Conference on Multimedia, MM ’09, 945–954. New York, NY, USA: Association for Computing Machinery. ISBN 9781605586083.
- Wang et al. (2022) Wang, J.; Yang, Z.; Hu, X.; Li, L.; Lin, K.; Gan, Z.; Liu, Z.; Liu, C.; and Wang, L. 2022. Git: A generative image-to-text transformer for vision and language. arXiv preprint arXiv:2205.14100.
- Wang et al. (2024) Wang, S.; Liu, S.; Tan, Z.; and Wang, X. 2024. Mindbridge: A cross-subject brain decoding framework. In Proceedings of the IEEE/CVF Conference on Computer Vision and Pattern Recognition, 11333–11342.
- Wang et al. (2004) Wang, Z.; Bovik, A. C.; Sheikh, H. R.; and Simoncelli, E. P. 2004. Image quality assessment: from error visibility to structural similarity. IEEE transactions on image processing, 13(4): 600–612.
- Wang, Murai, and Whitney (2020) Wang, Z.; Murai, Y.; and Whitney, D. 2020. Idiosyncratic perception: a link between acuity, perceived position and apparent size. Proceedings of the Royal Society B, 287(1930): 20200825.
- Xia et al. (2024a) Xia, W.; de Charette, R.; Oztireli, C.; and Xue, J.-H. 2024a. Dream: Visual decoding from reversing human visual system. In Proceedings of the IEEE/CVF Winter Conference on Applications of Computer Vision, 8226–8235.
- Xia et al. (2024b) Xia, W.; de Charette, R.; Öztireli, C.; and Xue, J.-H. 2024b. UMBRAE: Unified Multimodal Decoding of Brain Signals. arXiv preprint arXiv:2404.07202.
- Zhang, Rao, and Agrawala (2023) Zhang, L.; Rao, A.; and Agrawala, M. 2023. Adding conditional control to text-to-image diffusion models. In Proceedings of the IEEE/CVF International Conference on Computer Vision, 3836–3847.
- Zhao et al. (2014) Zhao, S.; Jiang, X.; Han, J.; Hu, X.; Zhu, D.; Lv, J.; Zhang, T.; Guo, L.; and Liu, T. 2014. Decoding Auditory Saliency from FMRI Brain Imaging. In Proceedings of the 22nd ACM International Conference on Multimedia, MM ’14, 873–876. New York, NY, USA: Association for Computing Machinery. ISBN 9781450330633.
Additional Details
Evaluation Metrics
Retrieval: Image retrieval refers to retrieving the image embeddings with the highest cosine similarity based on voxel embeddings on the test set (chance=0.3%). If a paired image embedding is retrieved, the retrieval is considered correct. Brain retrieval is the opposite process mentioned above.
PixCorr: pixel-wise correlation between ground truth and reconstructions;
SSIM: structural similarity index metric (Wang et al. 2004) between ground truth and reconstructions;
Eff (Tan and Le 2019)and Swav (Caron et al. 2020) refer to average correlation distance with EfficientNet-B1 and SwAV-ResNet50.
Alex(2), Alex(5), Incep, CLIP: all these metrics refer to two-way identification (chance = 50%) using different models. The two-way comparisons were performed with AlexNet where Alex(2) denotes the second layer, Alex(5) denotes the fifth layer, InceptionV3 with the last pooling layer, and CLIP with the final layer of ViT-L/14. Two-way identification refers to percent correct across comparisons gauging if the original image embedding is more similar to its paired voxel embedding or a randomly selected voxel embedding. We followed the same image preprocessing and two-way identification steps as (Ozcelik and VanRullen 2023; Scotti et al. 2023, 2024). For each test sample, performance was averaged across all possible pairwise comparisons using the other 999 reconstructions to ensure no bias from random sample selection. This yielded 1,000 averaged percent correct outputs.

Skip-LoRAs
As described in the main text, we designed Skip-LoRAs to learn better visual fingerprint between subjects, especially the non-linear part. The structure of Skip-LoRAs is shown in Figure 7. The pre-trained shared model of the MLP Backbone includes an alignment layer that aligns voxels to 4096-dim, four 4096-dim residual blocks, and a layer that maps to image tokens. We used Skip-LoRAs to connect fMRI to these layers, allowing the initial fMRI differences to affect the entire network. was also used for these connections, except for the last mapping layer of image tokens, as the non-linear constraints and retrieved cosine similarity-based contrastive loss conflicted with each other.
Additional Results
Adapative Porjector pre-trained on MSCOCO
Previous CLIP-related articles have demonstrated that ViT’s projection layer can take both CLS tokens and the result of the global average pooling of all tokens as inputs. We perform a global average pooling of all 256 brain tokens, which finally are mapped to the same 1280 dimensions as the CLS token of the text. The PyTorch code used to train the projector is depicted below:
We trained this projection layer on the whole COCO 2017 dataset including 73k images and 5 texts for each image. We randomly selected 900 image-text pairs as the validation set, and used a retrieval pool of 300 for validation (a similar setup to our main experiment). Instead of training from scratch, we used the projection layer of image CLS token from the open-source OpenCLIP333https://huggingface.co/laion/CLIP-ViT-bigG-14-laion2B-39B-b160k to start fine-tuning for 20 epochs, with the learning rate being fixed at 3e-4 and the batch size at 512. The loss is the standard CLIP image-text contrastive loss. The accuracy of validation set retrieval before and after fine-tuning is as follows:
Method | Image2Text | Text2Image |
---|---|---|
Before fine-tuning | 81.3% | 78.7% |
After fine-tuning | 95.2% | 95.2% |
Method | Data Size | Low-Level | High-Level | ||||||
---|---|---|---|---|---|---|---|---|---|
PixCorr | SSIM | Alex(2) | Alex(5) | Incep | CLIP | Eff | SwAV | ||
MindTuner(Uncorrected) | 40 hours | 0.280 | 0.327 | 95.4% | 99.2% | 96.4% | 94.5% | 0.621 | 0.341 |
MindTuner | 40 hours | 0.322 | 0.421 | 95.8% | 98.8% | 95.6% | 93.8% | 0.612 | 0.340 |
MindTuner(Uncorrected) | 1 hour | 0.208 | 0.378 | 87.4% | 94.0% | 85.5% | 83.8% | 0.781 | 0.439 |
MindTuner | 1 hour | 0.224 | 0.420 | 87.8% | 93.6% | 84.8% | 83.5% | 0.780 | 0.440 |
Semantic Correction Influence
As shown in Figure 5, we observe that semantic correction, while making the semantics of the reconstructed images clearer and more accurate, negatively affects the evaluation metrics, especially high-level. This is similar to MindEye2’s results, suggesting that manual correction of the image, either by MindEye2’s refinement or MindTuner’s correction, can have an adverse effect on the original fMRI representation. And as shown in Table 1, MindTuner mitigates this negative effect to some extent, compared to MindEye2.
Additional Brain Correlation Reults
The Brain Correlation results with 1-hour data are as follows. In conjunction with Table 3, it can be seen that MindTuner greatly improves the brain correlation of the reconstructed images, especially when the data is scarce. It suggests that the learning of subjects’ visual fingerprints preserves their own properties to some extent.
Brain Region | MindTuner | MindEye2 (Scotti et al. 2024) |
---|---|---|
Visual cortex | 0.372 | 0.348 |
V1 | 0.345 | 0.309 |
V2 | 0.348 | 0.314 |
V3 | 0.347 | 0.315 |
V4 | 0.329 | 0.300 |
Higher vis. | 0.370 | 0.351 |
1-hour data — | Subject 1 | Subject 2 | Subject 5 | Subject 7 | |
---|---|---|---|---|---|
Low-Level | PixCorr | 0.262 | 0.225 | 0.208 | 0.202 |
SSIM | 0.422 | 0.425 | 0.415 | 0.417 | |
Alex(2) | 90.6% | 89.1% | 86.8% | 84.5% | |
Alex(5) | 94.9% | 95.1% | 93.7% | 90.8% | |
High-Level | Incep | 85.8% | 84.8% | 87.7% | 80.7% |
CLIP | 84.6% | 83.7% | 85.9% | 79.6% | |
Eff | 0.774 | 0.781 | 0.750 | 0.817 | |
SwAV | 0.433 | 0.440 | 0.422 | 0.465 | |
Retrieval | Image | 94.2% | 93.7% | 72.2% | 71.9% |
Brain | 87.4% | 82.8% | 68.1% | 65.5% | |
Brain | Visual cortex | 0.359 | 0.376 | 0.426 | 0.328 |
V1 | 0.336 | 0.342 | 0.366 | 0.335 | |
Correlation | V2 | 0.354 | 0.326 | 0.373 | 0.339 |
V3 | 0.355 | 0.348 | 0.354 | 0.329 | |
V4 | 0.327 | 0.366 | 0.328 | 0.294 | |
Higher vis. | 0.358 | 0.379 | 0.432 | 0.311 |
40-hours data — | Subject 1 | Subject 2 | Subject 5 | Subject 7 | |
---|---|---|---|---|---|
Low-Level | PixCorr | 0.371 | 0.331 | 0.298 | 0.288 |
SSIM | 0.428 | 0.421 | 0.422 | 0.414 | |
Alex(2) | 97.7% | 96.8% | 94.7% | 94.1% | |
Alex(5) | 99.3% | 99.0% | 98.7% | 98.1% | |
High-Level | Incep | 96.4% | 95.1% | 96.6% | 94.1% |
CLIP | 94.3% | 92.7% | 94.7% | 93.3% | |
Eff | 0.601 | 0.620 | 0.592 | 0.636 | |
SwAV | 0.331 | 0.342 | 0.333 | 0.355 | |
Retrieval | Image | 100% | 99.9% | 98.4% | 97.2% |
Brain | 99.9% | 99.8% | 96.8% | 96.6% | |
Brain | Visual cortex | 0.377 | 0.394 | 0.416 | 0.320 |
V1 | 0.387 | 0.400 | 0.357 | 0.323 | |
Correlation | V2 | 0.381 | 0.359 | 0.361 | 0.317 |
V3 | 0.366 | 0.373 | 0.341 | 0.302 | |
V4 | 0.340 | 0.376 | 0.324 | 0.278 | |
Higher vis. | 0.363 | 0.387 | 0.426 | 0.311 |
Subject’s Specific Results
Tables 7 and 8 show more exhaustive evaluation metrics computed for every subject individually using 40-hours and 1-hour of fine-tuning data respectively. It can be seen that the performance of different subjects is relatively similar regardless of the amount of data, e.g., Subject 1 performs better at the low level and retrieval, while Subject 5 performs better at the high level.
Subject’s Specific Visualizations
We visualize in Figure 10 and Figure 11 the specific reconstruction results for different subjects, using either 1-hour or 40-hours data. For different subjects, MindTuner’s reconstructed images were accurate in capturing semantics, but the details varied. This may be attributed to the fact that the CLIP representation space is more semantic than detailed texture. In addition to this, as the training data increases, the blurry images become more similar to the ground truth images, thus giving a boost to the low-level metrics, as shown in Tables 10 and 11. Unfortunately, the color distribution of blurry images and reconstructed images is often difficult to control, e.g., generating airplanes of various colors. We have tried to add a sub-module to align the fMRI to a spatial palette, but the spatial palettes obtained are only slightly better than the blurry images. Moreover, the T2I Adapter generation model based on a spatial palette is difficult to control, which often makes the reconstruction results even worse. Thus, the relationship between the visual system and color needs to be further explored.
COCO Category
There are 80 categories of object categorization in MSCOCO, and the category text was used by us for semantic correction in MindTuner, below are the 80 categories:
’person, bicycle, car, motorcycle, airplane, bus, train, truck, boat, traffic light, fire hydrant, stop sign, parking meter, bench, bird, cat, dog, horse, sheep, cow, elephant, bear, zebra, giraffe, backpack, umbrella, handbag, tie, suitcase, frisbee, skis, snowboard, sports ball, kite, baseball bat, baseball glove, skateboard, surfboard, tennis racket, bottle, wine glass, cup, fork, knife, spoon, bowl, banana, apple, sandwich, orange, broccoli, carrot, hot dog, pizza, donut, cake, chair, couch, potted plant, bed, dining table, toilet, tv, laptop, mouse, remote, keyboard, cell phone, microwave, oven, toaster, sink, refrigerator, book, clock, vase, scissors, teddy bear, hair drier, toothbrush’
Failure Results
In this section, we visualize some of the failed reconstructed images, and we mainly identify semantically incorrect images as failed reconstructions because if the semantics are incorrect, it is even less important to talk about other details. All results shown in Figures 8 are from Subject 1.

The first reconstruction failures are purely due to insufficient training data, e.g., the bear in the first line and the bird in the second line have bad semantics when only 1 hour of training data is available, but the semantics are progressively more accurate over 40 hours. The second type of reconstruction failure is caused by similar categories resulting in insufficient semantic accuracy, e.g., bears and teddy bears, cakes and pizzas, etc. The third type of reconstruction failure is due to the inability of category nouns to describe overly complex scenarios, especially when a person is present. Since MSCOCO’s category noun is only ’person’, it often appears that the person determines the categorization result while ignoring other objects, and the gender of the person is often not accurately generated. This is because MindTuner only uses a single category noun, so more complex Pivot designs can be discussed in the future, whether for text generation or multi-classification.
Additional Neuroscience Visualizations
In this section, we visualize the voxel weight maps of the other three subjects in Figure 12, 13 and 14, and the templates for the brain region references associated with nsdgeneral are shown in Figure 9. It can be seen that the content of the visual fingerprint is not quite the same across subjects, but both LoRA and SKip-LoRA have a portion of the primary visual cortex voxels that are significantly more heavily weighted than Ridge. As with Subj01, The visual fingerprint is more heavily weighted on the higher visual cortex for Subj02 and Subj05. However, this phenomenon was not observed for Subj07, which may be the reason why Subj07 performs worse on all the metrics. Further reasons need more explanations from neuroscientific mechanisms.





