Mathematical and Scientific Machine Learning最新文献

筛选
英文 中文
SpecNet2: Orthogonalization-free spectral embedding by neural networks SpecNet2:基于神经网络的无正交化谱嵌入
Mathematical and Scientific Machine Learning Pub Date : 2022-06-14 DOI: 10.48550/arXiv.2206.06644
Ziyu Chen, Yingzhou Li, Xiuyuan Cheng
{"title":"SpecNet2: Orthogonalization-free spectral embedding by neural networks","authors":"Ziyu Chen, Yingzhou Li, Xiuyuan Cheng","doi":"10.48550/arXiv.2206.06644","DOIUrl":"https://doi.org/10.48550/arXiv.2206.06644","url":null,"abstract":"Spectral methods which represent data points by eigenvectors of kernel matrices or graph Laplacian matrices have been a primary tool in unsupervised data analysis. In many application scenarios, parametrizing the spectral embedding by a neural network that can be trained over batches of data samples gives a promising way to achieve automatic out-of-sample extension as well as computational scalability. Such an approach was taken in the original paper of SpectralNet (Shaham et al. 2018), which we call SpecNet1. The current paper introduces a new neural network approach, named SpecNet2, to compute spectral embedding which optimizes an equivalent objective of the eigen-problem and removes the orthogonalization layer in SpecNet1. SpecNet2 also allows separating the sampling of rows and columns of the graph affinity matrix by tracking the neighbors of each data point through the gradient formula. Theoretically, we show that any local minimizer of the new orthogonalization-free objective reveals the leading eigenvectors. Furthermore, global convergence for this new orthogonalization-free objective using a batch-based gradient descent method is proved. Numerical experiments demonstrate the improved performance and computational efficiency of SpecNet2 on simulated data and image datasets.","PeriodicalId":189279,"journal":{"name":"Mathematical and Scientific Machine Learning","volume":"43 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126772106","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Concentration of Random Feature Matrices in High-Dimensions 高维随机特征矩阵的集中
Mathematical and Scientific Machine Learning Pub Date : 2022-04-14 DOI: 10.48550/arXiv.2204.06935
Zhijun Chen, Hayden Schaeffer, Rachel A. Ward
{"title":"Concentration of Random Feature Matrices in High-Dimensions","authors":"Zhijun Chen, Hayden Schaeffer, Rachel A. Ward","doi":"10.48550/arXiv.2204.06935","DOIUrl":"https://doi.org/10.48550/arXiv.2204.06935","url":null,"abstract":"The spectra of random feature matrices provide essential information on the conditioning of the linear system used in random feature regression problems and are thus connected to the consistency and generalization of random feature models. Random feature matrices are asymmetric rectangular nonlinear matrices depending on two input variables, the data and the weights, which can make their characterization challenging. We consider two settings for the two input variables, either both are random variables or one is a random variable and the other is well-separated, i.e. there is a minimum distance between points. With conditions on the dimension, the complexity ratio, and the sampling variance, we show that the singular values of these matrices concentrate near their full expectation and near one with high-probability. In particular, since the dimension depends only on the logarithm of the number of random weights or the number of data points, our complexity bounds can be achieved even in moderate dimensions for many practical setting. The theoretical results are verified with numerical experiments.","PeriodicalId":189279,"journal":{"name":"Mathematical and Scientific Machine Learning","volume":"76 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"122611370","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Error-in-variables modelling for operator learning 算子学习的变量误差建模
Mathematical and Scientific Machine Learning Pub Date : 2022-04-01 DOI: 10.48550/arXiv.2204.10909
Ravi G. Patel, Indu Manickam, Myoungkyu Lee, Mamikon A. Gulian
{"title":"Error-in-variables modelling for operator learning","authors":"Ravi G. Patel, Indu Manickam, Myoungkyu Lee, Mamikon A. Gulian","doi":"10.48550/arXiv.2204.10909","DOIUrl":"https://doi.org/10.48550/arXiv.2204.10909","url":null,"abstract":"Deep operator learning has emerged as a promising tool for reduced-order modelling and PDE model discovery. Leveraging the expressive power of deep neural networks, especially in high dimensions, such methods learn the mapping between functional state variables. While proposed methods have assumed noise only in the dependent variables, experimental and numerical data for operator learning typically exhibit noise in the independent variables as well, since both variables represent signals that are subject to measurement error. In regression on scalar data, failure to account for noisy independent variables can lead to biased parameter estimates. With noisy independent variables, linear models fitted via ordinary least squares (OLS) will show attenuation bias, wherein the slope will be underestimated. In this work, we derive an analogue of attenuation bias for linear operator regression with white noise in both the independent and dependent variables. In the nonlinear setting, we computationally demonstrate underprediction of the action of the Burgers operator in the presence of noise in the independent variable. We propose error-in-variables (EiV) models for two operator regression methods, MOR-Physics and DeepONet, and demonstrate that these new models reduce bias in the presence of noisy independent variables for a variety of operator learning problems. Considering the Burgers operator in 1D and 2D, we demonstrate that EiV operator learning robustly recovers operators in high-noise regimes that defeat OLS operator learning. We also introduce an EiV model for time-evolving PDE discovery and show that OLS and EiV perform similarly in learning the Kuramoto-Sivashinsky evolution operator from corrupted data, suggesting that the effect of bias in OLS operator learning depends on the regularity of the target operator.","PeriodicalId":189279,"journal":{"name":"Mathematical and Scientific Machine Learning","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"123273371","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Monte Carlo Tree Search based Hybrid Optimization of Variational Quantum Circuits 基于蒙特卡罗树搜索的变分量子电路混合优化
Mathematical and Scientific Machine Learning Pub Date : 2022-03-30 DOI: 10.48550/arXiv.2203.16707
Jiahao Yao, Haoya Li, Marin Bukov, Lin Lin, Lexing Ying
{"title":"Monte Carlo Tree Search based Hybrid Optimization of Variational Quantum Circuits","authors":"Jiahao Yao, Haoya Li, Marin Bukov, Lin Lin, Lexing Ying","doi":"10.48550/arXiv.2203.16707","DOIUrl":"https://doi.org/10.48550/arXiv.2203.16707","url":null,"abstract":"Variational quantum algorithms stand at the forefront of simulations on near-term and future fault-tolerant quantum devices. While most variational quantum algorithms involve only continuous optimization variables, the representational power of the variational ansatz can sometimes be significantly enhanced by adding certain discrete optimization variables, as is exemplified by the generalized quantum approximate optimization algorithm (QAOA). However, the hybrid discrete-continuous optimization problem in the generalized QAOA poses a challenge to the optimization. We propose a new algorithm called MCTS-QAOA, which combines a Monte Carlo tree search method with an improved natural policy gradient solver to optimize the discrete and continuous variables in the quantum circuit, respectively. We find that MCTS-QAOA has excellent noise-resilience properties and outperforms prior algorithms in challenging instances of the generalized QAOA.","PeriodicalId":189279,"journal":{"name":"Mathematical and Scientific Machine Learning","volume":"19 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126232676","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Stochastic and Private Nonconvex Outlier-Robust PCA 随机和私有非凸离群值-鲁棒PCA
Mathematical and Scientific Machine Learning Pub Date : 2022-03-17 DOI: 10.48550/arXiv.2203.09276
Tyler Maunu, Chenyun Yu, Gilad Lerman
{"title":"Stochastic and Private Nonconvex Outlier-Robust PCA","authors":"Tyler Maunu, Chenyun Yu, Gilad Lerman","doi":"10.48550/arXiv.2203.09276","DOIUrl":"https://doi.org/10.48550/arXiv.2203.09276","url":null,"abstract":"We develop theoretically guaranteed stochastic methods for outlier-robust PCA. Outlier-robust PCA seeks an underlying low-dimensional linear subspace from a dataset that is corrupted with outliers. We are able to show that our methods, which involve stochastic geodesic gradient descent over the Grassmannian manifold, converge and recover an underlying subspace in various regimes through the development of a novel convergence analysis. The main application of this method is an effective differentially private algorithm for outlier-robust PCA that uses a Gaussian noise mechanism within the stochastic gradient method. Our results emphasize the advantages of the nonconvex methods over another convex approach to solving this problem in the differentially private setting. Experiments on synthetic and stylized data verify these results.","PeriodicalId":189279,"journal":{"name":"Mathematical and Scientific Machine Learning","volume":"47 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121273296","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Optimal denoising of rotationally invariant rectangular matrices 旋转不变矩形矩阵的最优去噪
Mathematical and Scientific Machine Learning Pub Date : 2022-03-15 DOI: 10.48550/arXiv.2203.07752
Emanuele Troiani, Vittorio Erba, F. Krzakala, Antoine Maillard, Lenka Zdeborov'a
{"title":"Optimal denoising of rotationally invariant rectangular matrices","authors":"Emanuele Troiani, Vittorio Erba, F. Krzakala, Antoine Maillard, Lenka Zdeborov'a","doi":"10.48550/arXiv.2203.07752","DOIUrl":"https://doi.org/10.48550/arXiv.2203.07752","url":null,"abstract":"In this manuscript we consider denoising of large rectangular matrices: given a noisy observation of a signal matrix, what is the best way of recovering the signal matrix itself? For Gaussian noise and rotationally-invariant signal priors, we completely characterize the optimal denoiser and its performance in the high-dimensional limit, in which the size of the signal matrix goes to infinity with fixed aspects ratio, and under the Bayes optimal setting, that is when the statistician knows how the signal and the observations were generated. Our results generalise previous works that considered only symmetric matrices to the more general case of non-symmetric and rectangular ones. We explore analytically and numerically a particular choice of factorized signal prior that models cross-covariance matrices and the matrix factorization problem. As a byproduct of our analysis, we provide an explicit asymptotic evaluation of the rectangular Harish-Chandra-Itzykson-Zuber integral in a special case.","PeriodicalId":189279,"journal":{"name":"Mathematical and Scientific Machine Learning","volume":"100 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"124830014","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
Data adaptive RKHS Tikhonov regularization for learning kernels in operators 算子核学习的数据自适应RKHS Tikhonov正则化
Mathematical and Scientific Machine Learning Pub Date : 2022-03-08 DOI: 10.48550/arXiv.2203.03791
F. Lu, Quanjun Lang, Qi An
{"title":"Data adaptive RKHS Tikhonov regularization for learning kernels in operators","authors":"F. Lu, Quanjun Lang, Qi An","doi":"10.48550/arXiv.2203.03791","DOIUrl":"https://doi.org/10.48550/arXiv.2203.03791","url":null,"abstract":"We present DARTR: a Data Adaptive RKHS Tikhonov Regularization method for the linear inverse problem of nonparametric learning of function parameters in operators. A key ingredient is a system intrinsic data-adaptive (SIDA) RKHS, whose norm restricts the learning to take place in the function space of identifiability. DARTR utilizes this norm and selects the regularization parameter by the L-curve method. We illustrate its performance in examples including integral operators, nonlinear operators and nonlocal operators with discrete synthetic data. Numerical results show that DARTR leads to an accurate estimator robust to both numerical error due to discrete data and noise in data, and the estimator converges at a consistent rate as the data mesh refines under different levels of noises, outperforming two baseline regularizers using $l^2$ and $L^2$ norms.","PeriodicalId":189279,"journal":{"name":"Mathematical and Scientific Machine Learning","volume":"253 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114465822","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Online Weak-form Sparse Identification of Partial Differential Equations 偏微分方程的在线弱形式稀疏辨识
Mathematical and Scientific Machine Learning Pub Date : 2022-03-08 DOI: 10.48550/arXiv.2203.03979
D. Messenger, E. Dall’Anese, D. Bortz
{"title":"Online Weak-form Sparse Identification of Partial Differential Equations","authors":"D. Messenger, E. Dall’Anese, D. Bortz","doi":"10.48550/arXiv.2203.03979","DOIUrl":"https://doi.org/10.48550/arXiv.2203.03979","url":null,"abstract":"This paper presents an online algorithm for identification of partial differential equations (PDEs) based on the weak-form sparse identification of nonlinear dynamics algorithm (WSINDy). The algorithm is online in a sense that if performs the identification task by processing solution snapshots that arrive sequentially. The core of the method combines a weak-form discretization of candidate PDEs with an online proximal gradient descent approach to the sparse regression problem. In particular, we do not regularize the $ell_0$-pseudo-norm, instead finding that directly applying its proximal operator (which corresponds to a hard thresholding) leads to efficient online system identification from noisy data. We demonstrate the success of the method on the Kuramoto-Sivashinsky equation, the nonlinear wave equation with time-varying wavespeed, and the linear wave equation, in one, two, and three spatial dimensions, respectively. In particular, our examples show that the method is capable of identifying and tracking systems with coefficients that vary abruptly in time, and offers a streaming alternative to problems in higher dimensions.","PeriodicalId":189279,"journal":{"name":"Mathematical and Scientific Machine Learning","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"129893957","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 9
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信