{"title":"Memory recall: Retrieval-Augmented mind reconstruction for brain decoding","authors":"Yuxiao Zhao , Guohua Dong , Lei Zhu , Xiaomin Ying","doi":"10.1016/j.inffus.2025.103280","DOIUrl":null,"url":null,"abstract":"<div><div>Reconstructing visual stimuli from functional magnetic resonance imaging (fMRI) is a complex challenge in neuroscience. Most existing approaches rely on mapping neural signals to pretrained models to generate latent variables, which are then used to reconstruct images via a diffusion model. However, this multi-step process can result in the loss of crucial semantic details, limiting reconstruction accuracy. In this paper, we introduce a novel brain decoding framework, called Memory Recall (MR), inspired by bionic brain mechanisms. MR mimics the human visual perception process, where the brain retrieves stored visual experiences to compensate for incomplete visual cues. Initially, low- and high-level visual cues are extracted using spatial mapping techniques based on VAE and CLIP, replicating the brain’s ability to interpret degraded stimuli. A visual experience database is then created to retrieve complementary information that enriches these high-level representations, simulating the brain’s memory retrieval process. Finally, an Attentive Visual Signal Fusion Network (AVSFN) with a novel attention scoring mechanism integrates the retrieved information, enhancing the generative model’s performance and emulating the brain’s refinement of visual perception. Experimental results show that MR outperforms state-of-the-art models across multiple evaluation metrics and subjective assessments. Moreover, our results provide new evidence supporting a well-known psychological conclusion that the basic information capacity of short-term memory is four items, further demonstrating the informativeness and interpretability of our model.</div></div>","PeriodicalId":50367,"journal":{"name":"Information Fusion","volume":"123 ","pages":"Article 103280"},"PeriodicalIF":14.7000,"publicationDate":"2025-05-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Information Fusion","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1566253525003537","RegionNum":1,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
引用次数: 0
Abstract
Reconstructing visual stimuli from functional magnetic resonance imaging (fMRI) is a complex challenge in neuroscience. Most existing approaches rely on mapping neural signals to pretrained models to generate latent variables, which are then used to reconstruct images via a diffusion model. However, this multi-step process can result in the loss of crucial semantic details, limiting reconstruction accuracy. In this paper, we introduce a novel brain decoding framework, called Memory Recall (MR), inspired by bionic brain mechanisms. MR mimics the human visual perception process, where the brain retrieves stored visual experiences to compensate for incomplete visual cues. Initially, low- and high-level visual cues are extracted using spatial mapping techniques based on VAE and CLIP, replicating the brain’s ability to interpret degraded stimuli. A visual experience database is then created to retrieve complementary information that enriches these high-level representations, simulating the brain’s memory retrieval process. Finally, an Attentive Visual Signal Fusion Network (AVSFN) with a novel attention scoring mechanism integrates the retrieved information, enhancing the generative model’s performance and emulating the brain’s refinement of visual perception. Experimental results show that MR outperforms state-of-the-art models across multiple evaluation metrics and subjective assessments. Moreover, our results provide new evidence supporting a well-known psychological conclusion that the basic information capacity of short-term memory is four items, further demonstrating the informativeness and interpretability of our model.
期刊介绍:
Information Fusion serves as a central platform for showcasing advancements in multi-sensor, multi-source, multi-process information fusion, fostering collaboration among diverse disciplines driving its progress. It is the leading outlet for sharing research and development in this field, focusing on architectures, algorithms, and applications. Papers dealing with fundamental theoretical analyses as well as those demonstrating their application to real-world problems will be welcome.