Hongxiao Wang, Yang Yang, Zhuo Zhao, Pengfei Gu, Nishchal Sapkota, Danny Z. Chen
{"title":"Path-GPTOmic: A Balanced Multi-modal Learning Framework for Survival Outcome Prediction","authors":"Hongxiao Wang, Yang Yang, Zhuo Zhao, Pengfei Gu, Nishchal Sapkota, Danny Z. Chen","doi":"arxiv-2403.11375","DOIUrl":null,"url":null,"abstract":"For predicting cancer survival outcomes, standard approaches in clinical\nresearch are often based on two main modalities: pathology images for observing\ncell morphology features, and genomic (e.g., bulk RNA-seq) for quantifying gene\nexpressions. However, existing pathology-genomic multi-modal algorithms face\nsignificant challenges: (1) Valuable biological insights regarding genes and\ngene-gene interactions are frequently overlooked; (2) one modality often\ndominates the optimization process, causing inadequate training for the other\nmodality. In this paper, we introduce a new multi-modal ``Path-GPTOmic\"\nframework for cancer survival outcome prediction. First, to extract valuable\nbiological insights, we regulate the embedding space of a foundation model,\nscGPT, initially trained on single-cell RNA-seq data, making it adaptable for\nbulk RNA-seq data. Second, to address the imbalance-between-modalities problem,\nwe propose a gradient modulation mechanism tailored to the Cox partial\nlikelihood loss for survival prediction. The contributions of the modalities\nare dynamically monitored and adjusted during the training process, encouraging\nthat both modalities are sufficiently trained. Evaluated on two TCGA(The Cancer\nGenome Atlas) datasets, our model achieves substantially improved survival\nprediction accuracy.","PeriodicalId":501070,"journal":{"name":"arXiv - QuanBio - Genomics","volume":"31 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-03-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuanBio - Genomics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2403.11375","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
For predicting cancer survival outcomes, standard approaches in clinical
research are often based on two main modalities: pathology images for observing
cell morphology features, and genomic (e.g., bulk RNA-seq) for quantifying gene
expressions. However, existing pathology-genomic multi-modal algorithms face
significant challenges: (1) Valuable biological insights regarding genes and
gene-gene interactions are frequently overlooked; (2) one modality often
dominates the optimization process, causing inadequate training for the other
modality. In this paper, we introduce a new multi-modal ``Path-GPTOmic"
framework for cancer survival outcome prediction. First, to extract valuable
biological insights, we regulate the embedding space of a foundation model,
scGPT, initially trained on single-cell RNA-seq data, making it adaptable for
bulk RNA-seq data. Second, to address the imbalance-between-modalities problem,
we propose a gradient modulation mechanism tailored to the Cox partial
likelihood loss for survival prediction. The contributions of the modalities
are dynamically monitored and adjusted during the training process, encouraging
that both modalities are sufficiently trained. Evaluated on two TCGA(The Cancer
Genome Atlas) datasets, our model achieves substantially improved survival
prediction accuracy.