Wei Guo, Fuzhen Zhuang, Xiao Zhang, Yiqi Tong, Jin Dong
{"title":"A comprehensive survey of federated transfer learning: challenges, methods and applications","authors":"Wei Guo, Fuzhen Zhuang, Xiao Zhang, Yiqi Tong, Jin Dong","doi":"10.1007/s11704-024-40065-x","DOIUrl":"https://doi.org/10.1007/s11704-024-40065-x","url":null,"abstract":"<p>Federated learning (FL) is a novel distributed machine learning paradigm that enables participants to collaboratively train a centralized model with privacy preservation by eliminating the requirement of data sharing. In practice, FL often involves multiple participants and requires the third party to aggregate global information to guide the update of the target participant. Therefore, many FL methods do not work well due to the training and test data of each participant may not be sampled from the same feature space and the same underlying distribution. Meanwhile, the differences in their local devices (system heterogeneity), the continuous influx of online data (incremental data), and labeled data scarcity may further influence the performance of these methods. To solve this problem, federated transfer learning (FTL), which integrates transfer learning (TL) into FL, has attracted the attention of numerous researchers. However, since FL enables a continuous share of knowledge among participants with each communication round while not allowing local data to be accessed by other participants, FTL faces many unique challenges that are not present in TL. In this survey, we focus on categorizing and reviewing the current progress on federated transfer learning, and outlining corresponding solutions and applications. Furthermore, the common setting of FTL scenarios, available datasets, and significant related research are summarized in this survey.</p>","PeriodicalId":12640,"journal":{"name":"Frontiers of Computer Science","volume":"93 1","pages":""},"PeriodicalIF":4.2,"publicationDate":"2024-07-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141770853","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Graph foundation model","authors":"Chuan Shi, Junze Chen, Jiawei Liu, Cheng Yang","doi":"10.1007/s11704-024-40046-0","DOIUrl":"https://doi.org/10.1007/s11704-024-40046-0","url":null,"abstract":"<p>Graph Foundation Models represent an evolving direction in graph machine learning. Drawing inspiration from the success of Large Language Models in NLP, GFMs are designed to be trained on extensive graph data and adapted for a diverse array of downstream tasks. In this article, we have explained and introduced the concept of GFMs, comparing them with Language Foundation Models to highlight their similarities and differences. We identified the key technologies in building GFMs as the pre-train and adaptation techniques from the fields of GNNs and LLMs. Additionally, we discussed the potential for GFMs to have significant applications in various domains, ranging from social network analysis to bioinformatics and beyond.</p>","PeriodicalId":12640,"journal":{"name":"Frontiers of Computer Science","volume":"16 1","pages":""},"PeriodicalIF":4.2,"publicationDate":"2024-07-05","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141549372","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"SEOE: an option graph based semantically embedding method for prenatal depression detection","authors":"Xiaosong Han, Mengchen Cao, Dong Xu, Xiaoyue Feng, Yanchun Liang, Xiaoduo Lang, Renchu Guan","doi":"10.1007/s11704-024-3612-4","DOIUrl":"https://doi.org/10.1007/s11704-024-3612-4","url":null,"abstract":"<p>Prenatal depression, which can affect pregnant women’s physical and psychological health and cause postpartum depression, is increasing dramatically. Therefore, it is essential to detect prenatal depression early and conduct an attribution analysis. Many studies have used questionnaires to screen for prenatal depression, but the existing methods lack attributability. To diagnose the early signs of prenatal depression and identify the key factors that may lead to prenatal depression from questionnaires, we present the semantically enhanced option embedding (SEOE) model to represent questionnaire options. It can quantitatively determine the relationship and patterns between options and depression. SEOE first quantifies options and resorts them, gathering options with little difference, since Word2Vec is highly dependent on context. The resort task is transformed into an optimization problem involving the traveling salesman problem. Moreover, all questionnaire samples are used to train the options’ vector using Word2Vec. Finally, an LSTM and GRU fused model incorporating the cycle learning rate is constructed to detect whether a pregnant woman is suffering from depression. To verify the model, we compare it with other deep learning and traditional machine learning methods. The experiment results show that our proposed model can accurately identify pregnant women with depression and reach an F1 score of 0.8. The most relevant factors of depression found by SEOE are also verified in the literature. In addition, our model is of low computational complexity and strong generalization, which can be widely applied to other questionnaire analyses of psychiatric disorders.</p>","PeriodicalId":12640,"journal":{"name":"Frontiers of Computer Science","volume":"3 1","pages":""},"PeriodicalIF":4.2,"publicationDate":"2024-06-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141501822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"WPIA: accelerating DNN warm-up in Web browsers by precompiling WebGL programs","authors":"Deyu Tian, Yun Ma, Yudong Han, Qi Yang, Haochen Yang, Gang Huang","doi":"10.1007/s11704-024-40066-w","DOIUrl":"https://doi.org/10.1007/s11704-024-40066-w","url":null,"abstract":"<p>In this paper, we study the long warm-up time of GPU acceleration of DNN inference in Web browsers. We analyzed the reason behind the long warm-up time through a measurement study and revealed that compiling WebGL programs takes most of the warm-up time. Inspired by this finding, we proposed WPIA, an approach that suggests precompiling WebGL programs on the server side to avoid compiling them in Web browsers. WPIA tackles the challenges of precompiling by merging WebGL programs and using a record-and-replay technique. Evaluation experiment results show that WPIA can accelerate the DNN warm-up time to an order of magnitude.</p>","PeriodicalId":12640,"journal":{"name":"Frontiers of Computer Science","volume":"17 1","pages":""},"PeriodicalIF":4.2,"publicationDate":"2024-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141501824","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"Audio-guided self-supervised learning for disentangled visual speech representations","authors":"Dalu Feng, Shuang Yang, Shiguang Shan, Xilin Chen","doi":"10.1007/s11704-024-3787-8","DOIUrl":"https://doi.org/10.1007/s11704-024-3787-8","url":null,"abstract":"<p>In this paper, we propose a novel two-branch framework to learn the disentangled visual speech representations based on two particular observations. Its main idea is to introduce the audio signal to guide the learning of speech-relevant cues and introduce a bottleneck to restrict the speech-irrelevant branch from learning high-frequency and fine-grained speech cues. Experiments on both the word-level and sentence-level audio-visual speech datasets LRW and LRS2-BBC show the effectiveness. Our future work is to explore more explicit auxiliary tasks and constraints beyond the reconstruction task of the speech-relevant and irrelevant branch to improve further its ability of capturing speech cues in the video. Meanwhile, it’s also a nice try to combine multiple types of knowledge representations [10] to further boost the obtained speech epresentations, which is also left for the future work.</p>","PeriodicalId":12640,"journal":{"name":"Frontiers of Computer Science","volume":"75 1","pages":""},"PeriodicalIF":4.2,"publicationDate":"2024-06-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141501825","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Da-Wei Zhou, Zhi-Hong Qi, Han-Jia Ye, De-Chuan Zhan
{"title":"TV100: a TV series dataset that pre-trained CLIP has not seen","authors":"Da-Wei Zhou, Zhi-Hong Qi, Han-Jia Ye, De-Chuan Zhan","doi":"10.1007/s11704-024-40217-z","DOIUrl":"https://doi.org/10.1007/s11704-024-40217-z","url":null,"abstract":"<p>The era of pre-trained models has ushered in a wealth of new insights for the machine learning community. Among the myriad of questions that arise, one of paramount importance is: ‘Do pre-trained models possess comprehensive knowledge?’ This paper seeks to address this crucial inquiry. In line with our objective, we have made publicly available a novel dataset comprised of images from TV series released post-2021. This dataset holds significant potential for use in various research areas, including the evaluation of novel class iscovery and long-tailed learning, among others.</p>","PeriodicalId":12640,"journal":{"name":"Frontiers of Computer Science","volume":"128 1","pages":""},"PeriodicalIF":4.2,"publicationDate":"2024-06-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"141552704","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"HeterMM: applying in-DRAM index to heterogeneous memory-based key-value stores","authors":"Yunhong Ji, Wentao Huang, Xuan Zhou","doi":"10.1007/s11704-024-3713-0","DOIUrl":"https://doi.org/10.1007/s11704-024-3713-0","url":null,"abstract":"<p>We propose HeterMM, a versatile framework that leverages in-DRAM indexes in KV stores on heterogeneous memory. HeterMM incorporates a plug-in programming model, allowing for the integration of various types of indexes. By prioritizing the maintenance of both index and hot data in DRAM, HeterMM maximizes the utilization of the superior performance of DRAM. Our evaluation demonstrates that HeterMM outperforms existing state-of-the-art frameworks that convert in-DRAM indexes to persistent ones. Furthermore, HeterMM can surpass NVM-specific KV stores by carefully selecting the appropriate index for specific scenarios.</p>","PeriodicalId":12640,"journal":{"name":"Frontiers of Computer Science","volume":"36 1","pages":""},"PeriodicalIF":4.2,"publicationDate":"2024-04-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140586741","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
{"title":"A survey on large language model based autonomous agents","authors":"Lei Wang, Chen Ma, Xueyang Feng, Zeyu Zhang, Hao Yang, Jingsen Zhang, Zhiyuan Chen, Jiakai Tang, Xu Chen, Yankai Lin, Wayne Xin Zhao, Zhewei Wei, Jirong Wen","doi":"10.1007/s11704-024-40231-1","DOIUrl":"https://doi.org/10.1007/s11704-024-40231-1","url":null,"abstract":"<p>Autonomous agents have long been a research focus in academic and industry communities. Previous research often focuses on training agents with limited knowledge within isolated environments, which diverges significantly from human learning processes, and makes the agents hard to achieve human-like decisions. Recently, through the acquisition of vast amounts of Web knowledge, large language models (LLMs) have shown potential in human-level intelligence, leading to a surge in research on LLM-based autonomous agents. In this paper, we present a comprehensive survey of these studies, delivering a systematic review of LLM-based autonomous agents from a holistic perspective. We first discuss the construction of LLM-based autonomous agents, proposing a unified framework that encompasses much of previous work. Then, we present a overview of the diverse applications of LLM-based autonomous agents in social science, natural science, and engineering. Finally, we delve into the evaluation strategies commonly used for LLM-based autonomous agents. Based on the previous studies, we also present several challenges and future directions in this field.</p>","PeriodicalId":12640,"journal":{"name":"Frontiers of Computer Science","volume":"304 1","pages":""},"PeriodicalIF":4.2,"publicationDate":"2024-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140199133","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Haixin Wang, Yunhan Wang, Qun Jiang, Yan Zhang, Shengquan Chen
{"title":"SCREEN: predicting single-cell gene expression perturbation responses via optimal transport","authors":"Haixin Wang, Yunhan Wang, Qun Jiang, Yan Zhang, Shengquan Chen","doi":"10.1007/s11704-024-31014-9","DOIUrl":"https://doi.org/10.1007/s11704-024-31014-9","url":null,"abstract":"<p>In this study, we propose SCREEN, a novel method for predicting perturbation responses of scRNA-seq data. Through extensive experiments on various datasets, we validated the effectiveness and advantages of SCREEN for the prediction of single-cell gene expression perturbation responses. Besides, we demonstrated the ability of SCREEN to facilitate biological implications in downstream analysis. Moreover, we showed the robustness of SCREEN to noise degree, number of cell types, and cell type imbalance, indicating its broader applicability. Source codes and detailed tutorials of SCREEN are freely accessible at Github (Califorya/SCREEN). We anticipate SCREEN will greatly assist with perturbational single-cell omics and precision medicine.</p>","PeriodicalId":12640,"journal":{"name":"Frontiers of Computer Science","volume":"13 1","pages":""},"PeriodicalIF":4.2,"publicationDate":"2024-03-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140199135","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
Yuqi Li, Tao Meng, Zhixiong He, Haiyan Liu, Keqin Li
{"title":"A biased edge enhancement method for truss-based community search","authors":"Yuqi Li, Tao Meng, Zhixiong He, Haiyan Liu, Keqin Li","doi":"10.1007/s11704-024-2604-8","DOIUrl":"https://doi.org/10.1007/s11704-024-2604-8","url":null,"abstract":"<p>Most truss-based community search methods are usually confronted with the fragmentation issue. We propose a Biased edge Enhancement method for Truss-based Community Search (BETCS) to address the issue. This paper mainly solves the fragmentation problem in truss community query through data enhancement. In future work, we will consider applying the methods in the text to directed graphs or dynamic graphs.</p>","PeriodicalId":12640,"journal":{"name":"Frontiers of Computer Science","volume":"16 1","pages":""},"PeriodicalIF":4.2,"publicationDate":"2024-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"140149111","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}