{"title":"Two-View and Multi-View Tensor Canonical Correlation Analysis Over Graphs","authors":"Thummaluru Siddartha Reddy;Sundeep Prabhakar Chepuri","doi":"10.1109/TSIPN.2025.3536102","DOIUrl":null,"url":null,"abstract":"In this work, we focus on the multi-view dimensionality reduction problem for tensor data on graphs. In particular, we extend canonical correlation analysis on graphs <monospace>(CCA-G)</monospace> and multi-view canonical correlation analysis on graphs <monospace>(MCCA-G)</monospace> for tensor data. Directly applying <monospace>CCA-G</monospace> and <monospace>MCCA-G</monospace> to tensor data requires vectorization, which destroys the underlying structure in the data and often outputs very high-dimensional data leading to the curse of dimensionality. To circumvent the vectorization operation, we propose tensor canonical correlation analysis on graphs <monospace>(TCCA-G)</monospace> for two view data and tensor multi-view canonical correlation analysis on graphs <monospace>(TMCCA-G)</monospace> for multi-view tensor data that preserves the intrinsic structure in data and accounts for underlying graph structure in the latent variable. In particular, the proposed <monospace>TCCA-G</monospace> promotes smoothness of the tensor canonical variates over a graph and outputs the tensor canonical variates that are correlated within the set and uncorrelated across the sets. In the absence of prior (smoothness) information on the latent variable, <monospace>TCCA-G</monospace> simplifies to tensor canonical correlation analysis <monospace>(TCCA)</monospace> that only preserves the intrinsic structure in the data and results in an uncorrelated set of features. To solve <monospace>TCCA-G</monospace> and <monospace>TCCA</monospace>, we present an algorithm based on alternating minimization. In particular, the canonical subspaces in <monospace>TCCA</monospace> and <monospace>TCCA-G</monospace> are obtained by solving an eigenvalue problem. <monospace>TMCCA-G</monospace> extends <monospace>TCCA-G</monospace> to multi-view data, wherein <monospace>TMCCA-G</monospace> obtains the canonical subspaces by solving a simple least-squares problem and the common source is obtained recursively using a Crank-Nicolson-like update to preserve the orthonormality constraints. In the absence of the graph prior, we present tensor multi-view canonical correlation analysis <monospace>(TMCCA)</monospace>, in which the common source is obtained in closed-form by solving an orthogonal Procustes problem. Therefore, each subproblem in <monospace>TMCCA</monospace> admits a closed-form solution in contrast to <monospace>TMCCA-G</monospace>. We show the efficacy of proposed algorithms through experiments on diverse tasks such as classification and clustering on real datasets.","PeriodicalId":56268,"journal":{"name":"IEEE Transactions on Signal and Information Processing over Networks","volume":"11 ","pages":"535-550"},"PeriodicalIF":3.0000,"publicationDate":"2025-01-29","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Signal and Information Processing over Networks","FirstCategoryId":"94","ListUrlMain":"https://ieeexplore.ieee.org/document/10857651/","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
In this work, we focus on the multi-view dimensionality reduction problem for tensor data on graphs. In particular, we extend canonical correlation analysis on graphs (CCA-G) and multi-view canonical correlation analysis on graphs (MCCA-G) for tensor data. Directly applying CCA-G and MCCA-G to tensor data requires vectorization, which destroys the underlying structure in the data and often outputs very high-dimensional data leading to the curse of dimensionality. To circumvent the vectorization operation, we propose tensor canonical correlation analysis on graphs (TCCA-G) for two view data and tensor multi-view canonical correlation analysis on graphs (TMCCA-G) for multi-view tensor data that preserves the intrinsic structure in data and accounts for underlying graph structure in the latent variable. In particular, the proposed TCCA-G promotes smoothness of the tensor canonical variates over a graph and outputs the tensor canonical variates that are correlated within the set and uncorrelated across the sets. In the absence of prior (smoothness) information on the latent variable, TCCA-G simplifies to tensor canonical correlation analysis (TCCA) that only preserves the intrinsic structure in the data and results in an uncorrelated set of features. To solve TCCA-G and TCCA, we present an algorithm based on alternating minimization. In particular, the canonical subspaces in TCCA and TCCA-G are obtained by solving an eigenvalue problem. TMCCA-G extends TCCA-G to multi-view data, wherein TMCCA-G obtains the canonical subspaces by solving a simple least-squares problem and the common source is obtained recursively using a Crank-Nicolson-like update to preserve the orthonormality constraints. In the absence of the graph prior, we present tensor multi-view canonical correlation analysis (TMCCA), in which the common source is obtained in closed-form by solving an orthogonal Procustes problem. Therefore, each subproblem in TMCCA admits a closed-form solution in contrast to TMCCA-G. We show the efficacy of proposed algorithms through experiments on diverse tasks such as classification and clustering on real datasets.
期刊介绍:
The IEEE Transactions on Signal and Information Processing over Networks publishes high-quality papers that extend the classical notions of processing of signals defined over vector spaces (e.g. time and space) to processing of signals and information (data) defined over networks, potentially dynamically varying. In signal processing over networks, the topology of the network may define structural relationships in the data, or may constrain processing of the data. Topics include distributed algorithms for filtering, detection, estimation, adaptation and learning, model selection, data fusion, and diffusion or evolution of information over such networks, and applications of distributed signal processing.