arXiv - STAT - Machine Learning最新文献

筛选
英文 中文
A clustering adaptive Gaussian process regression method: response patterns based real-time prediction for nonlinear solid mechanics problems 聚类自适应高斯过程回归方法:基于响应模式的非线性固体力学问题实时预测
arXiv - STAT - Machine Learning Pub Date : 2024-09-15 DOI: arxiv-2409.10572
Ming-Jian Li, Yanping Lian, Zhanshan Cheng, Lehui Li, Zhidong Wang, Ruxin Gao, Daining Fang
{"title":"A clustering adaptive Gaussian process regression method: response patterns based real-time prediction for nonlinear solid mechanics problems","authors":"Ming-Jian Li, Yanping Lian, Zhanshan Cheng, Lehui Li, Zhidong Wang, Ruxin Gao, Daining Fang","doi":"arxiv-2409.10572","DOIUrl":"https://doi.org/arxiv-2409.10572","url":null,"abstract":"Numerical simulation is powerful to study nonlinear solid mechanics problems.\u0000However, mesh-based or particle-based numerical methods suffer from the common\u0000shortcoming of being time-consuming, particularly for complex problems with\u0000real-time analysis requirements. This study presents a clustering adaptive\u0000Gaussian process regression (CAG) method aiming for real-time prediction for\u0000nonlinear structural responses in solid mechanics. It is a data-driven machine\u0000learning method featuring a small sample size, high accuracy, and high\u0000efficiency, leveraging nonlinear structural response patterns. Similar to the\u0000traditional Gaussian process regression (GPR) method, it operates in offline\u0000and online stages. In the offline stage, an adaptive sample generation\u0000technique is introduced to cluster datasets into distinct patterns for\u0000demand-driven sample allocation. This ensures comprehensive coverage of the\u0000critical samples for the solution space of interest. In the online stage,\u0000following the divide-and-conquer strategy, a pre-prediction classification\u0000categorizes problems into predefined patterns sequentially predicted by the\u0000trained multi-pattern Gaussian process regressor. In addition, dimension\u0000reduction and restoration techniques are employed in the proposed method to\u0000enhance its efficiency. A set of problems involving material, geometric, and\u0000boundary condition nonlinearities is presented to demonstrate the CAG method's\u0000abilities. The proposed method can offer predictions within a second and attain\u0000high precision with only about 20 samples within the context of this study,\u0000outperforming the traditional GPR using uniformly distributed samples for error\u0000reductions ranging from 1 to 3 orders of magnitude. The CAG method is expected\u0000to offer a powerful tool for real-time prediction of nonlinear solid mechanical\u0000problems and shed light on the complex nonlinear structural response pattern.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261682","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Consistent Spectral Clustering in Hyperbolic Spaces 双曲空间中的一致谱聚类
arXiv - STAT - Machine Learning Pub Date : 2024-09-14 DOI: arxiv-2409.09304
Sagar Ghosh, Swagatam Das
{"title":"Consistent Spectral Clustering in Hyperbolic Spaces","authors":"Sagar Ghosh, Swagatam Das","doi":"arxiv-2409.09304","DOIUrl":"https://doi.org/arxiv-2409.09304","url":null,"abstract":"Clustering, as an unsupervised technique, plays a pivotal role in various\u0000data analysis applications. Among clustering algorithms, Spectral Clustering on\u0000Euclidean Spaces has been extensively studied. However, with the rapid\u0000evolution of data complexity, Euclidean Space is proving to be inefficient for\u0000representing and learning algorithms. Although Deep Neural Networks on\u0000hyperbolic spaces have gained recent traction, clustering algorithms or\u0000non-deep machine learning models on non-Euclidean Spaces remain underexplored.\u0000In this paper, we propose a spectral clustering algorithm on Hyperbolic Spaces\u0000to address this gap. Hyperbolic Spaces offer advantages in representing complex\u0000data structures like hierarchical and tree-like structures, which cannot be\u0000embedded efficiently in Euclidean Spaces. Our proposed algorithm replaces the\u0000Euclidean Similarity Matrix with an appropriate Hyperbolic Similarity Matrix,\u0000demonstrating improved efficiency compared to clustering in Euclidean Spaces.\u0000Our contributions include the development of the spectral clustering algorithm\u0000on Hyperbolic Spaces and the proof of its weak consistency. We show that our\u0000algorithm converges at least as fast as Spectral Clustering on Euclidean\u0000Spaces. To illustrate the efficacy of our approach, we present experimental\u0000results on the Wisconsin Breast Cancer Dataset, highlighting the superior\u0000performance of Hyperbolic Spectral Clustering over its Euclidean counterpart.\u0000This work opens up avenues for utilizing non-Euclidean Spaces in clustering\u0000algorithms, offering new perspectives for handling complex data structures and\u0000improving clustering efficiency.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261749","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Active Learning to Guide Labeling Efforts for Question Difficulty Estimation 通过主动学习引导问题难度估算的标记工作
arXiv - STAT - Machine Learning Pub Date : 2024-09-14 DOI: arxiv-2409.09258
Arthur Thuy, Ekaterina Loginova, Dries F. Benoit
{"title":"Active Learning to Guide Labeling Efforts for Question Difficulty Estimation","authors":"Arthur Thuy, Ekaterina Loginova, Dries F. Benoit","doi":"arxiv-2409.09258","DOIUrl":"https://doi.org/arxiv-2409.09258","url":null,"abstract":"In recent years, there has been a surge in research on Question Difficulty\u0000Estimation (QDE) using natural language processing techniques.\u0000Transformer-based neural networks achieve state-of-the-art performance,\u0000primarily through supervised methods but with an isolated study in unsupervised\u0000learning. While supervised methods focus on predictive performance, they\u0000require abundant labeled data. On the other hand, unsupervised methods do not\u0000require labeled data but rely on a different evaluation metric that is also\u0000computationally expensive in practice. This work bridges the research gap by\u0000exploring active learning for QDE, a supervised human-in-the-loop approach\u0000striving to minimize the labeling efforts while matching the performance of\u0000state-of-the-art models. The active learning process iteratively trains on a\u0000labeled subset, acquiring labels from human experts only for the most\u0000informative unlabeled data points. Furthermore, we propose a novel acquisition\u0000function PowerVariance to add the most informative samples to the labeled set,\u0000a regression extension to the PowerBALD function popular in classification. We\u0000employ DistilBERT for QDE and identify informative samples by applying Monte\u0000Carlo dropout to capture epistemic uncertainty in unlabeled samples. The\u0000experiments demonstrate that active learning with PowerVariance acquisition\u0000achieves a performance close to fully supervised models after labeling only 10%\u0000of the training data. The proposed methodology promotes the responsible use of\u0000educational resources, makes QDE tools more accessible to course instructors,\u0000and is promising for other applications such as personalized support systems\u0000and question-answering tools.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261753","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
BM$^2$: Coupled Schrödinger Bridge Matching BM$^2$:耦合薛定谔桥匹配
arXiv - STAT - Machine Learning Pub Date : 2024-09-14 DOI: arxiv-2409.09376
Stefano Peluchetti
{"title":"BM$^2$: Coupled Schrödinger Bridge Matching","authors":"Stefano Peluchetti","doi":"arxiv-2409.09376","DOIUrl":"https://doi.org/arxiv-2409.09376","url":null,"abstract":"A Schr\"{o}dinger bridge establishes a dynamic transport map between two\u0000target distributions via a reference process, simultaneously solving an\u0000associated entropic optimal transport problem. We consider the setting where\u0000samples from the target distributions are available, and the reference\u0000diffusion process admits tractable dynamics. We thus introduce Coupled Bridge\u0000Matching (BM$^2$), a simple emph{non-iterative} approach for learning\u0000Schr\"{o}dinger bridges with neural networks. A preliminary theoretical\u0000analysis of the convergence properties of BM$^2$ is carried out, supported by\u0000numerical experiments that demonstrate the effectiveness of our proposal.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261746","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Beta-Sigma VAE: Separating beta and decoder variance in Gaussian variational autoencoder 贝塔-西格玛 VAE:分离高斯变异自动编码器中的贝塔方差和解码器方差
arXiv - STAT - Machine Learning Pub Date : 2024-09-14 DOI: arxiv-2409.09361
Seunghwan Kim, Seungkyu Lee
{"title":"Beta-Sigma VAE: Separating beta and decoder variance in Gaussian variational autoencoder","authors":"Seunghwan Kim, Seungkyu Lee","doi":"arxiv-2409.09361","DOIUrl":"https://doi.org/arxiv-2409.09361","url":null,"abstract":"Variational autoencoder (VAE) is an established generative model but is\u0000notorious for its blurriness. In this work, we investigate the blurry output\u0000problem of VAE and resolve it, exploiting the variance of Gaussian decoder and\u0000$beta$ of beta-VAE. Specifically, we reveal that the indistinguishability of\u0000decoder variance and $beta$ hinders appropriate analysis of the model by\u0000random likelihood value, and limits performance improvement by omitting the\u0000gain from $beta$. To address the problem, we propose Beta-Sigma VAE (BS-VAE)\u0000that explicitly separates $beta$ and decoder variance $sigma^2_x$ in the\u0000model. Our method demonstrates not only superior performance in natural image\u0000synthesis but also controllable parameters and predictable analysis compared to\u0000conventional VAE. In our experimental evaluation, we employ the analysis of\u0000rate-distortion curve and proxy metrics on computer vision datasets. The code\u0000is available on https://github.com/overnap/BS-VAE","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261747","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Topological Tensor Eigenvalue Theorems in Data Fusion 数据融合中的拓扑张量特征值定理
arXiv - STAT - Machine Learning Pub Date : 2024-09-14 DOI: arxiv-2409.09392
Ronald Katende
{"title":"Topological Tensor Eigenvalue Theorems in Data Fusion","authors":"Ronald Katende","doi":"arxiv-2409.09392","DOIUrl":"https://doi.org/arxiv-2409.09392","url":null,"abstract":"This paper introduces a novel framework for tensor eigenvalue analysis in the\u0000context of multi-modal data fusion, leveraging topological invariants such as\u0000Betti numbers. While traditional approaches to tensor eigenvalues rely on\u0000algebraic extensions of matrix theory, this work provides a topological\u0000perspective that enriches the understanding of tensor structures. By\u0000establishing new theorems linking eigenvalues to topological features, the\u0000proposed framework offers deeper insights into the latent structure of data,\u0000enhancing both interpretability and robustness. Applications to data fusion\u0000illustrate the theoretical and practical significance of the approach,\u0000demonstrating its potential for broad impact across machine learning and data\u0000science domains.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Schrödinger Bridge Flow for Unpaired Data Translation 用于非配对数据转换的薛定谔桥流
arXiv - STAT - Machine Learning Pub Date : 2024-09-14 DOI: arxiv-2409.09347
Valentin De Bortoli, Iryna Korshunova, Andriy Mnih, Arnaud Doucet
{"title":"Schrödinger Bridge Flow for Unpaired Data Translation","authors":"Valentin De Bortoli, Iryna Korshunova, Andriy Mnih, Arnaud Doucet","doi":"arxiv-2409.09347","DOIUrl":"https://doi.org/arxiv-2409.09347","url":null,"abstract":"Mass transport problems arise in many areas of machine learning whereby one\u0000wants to compute a map transporting one distribution to another. Generative\u0000modeling techniques like Generative Adversarial Networks (GANs) and Denoising\u0000Diffusion Models (DDMs) have been successfully adapted to solve such transport\u0000problems, resulting in CycleGAN and Bridge Matching respectively. However,\u0000these methods do not approximate Optimal Transport (OT) maps, which are known\u0000to have desirable properties. Existing techniques approximating OT maps for\u0000high-dimensional data-rich problems, such as DDM-based Rectified Flow and\u0000Schr\"odinger Bridge procedures, require fully training a DDM-type model at\u0000each iteration, or use mini-batch techniques which can introduce significant\u0000errors. We propose a novel algorithm to compute the Schr\"odinger Bridge, a\u0000dynamic entropy-regularised version of OT, that eliminates the need to train\u0000multiple DDM-like models. This algorithm corresponds to a discretisation of a\u0000flow of path measures, which we call the Schr\"odinger Bridge Flow, whose only\u0000stationary point is the Schr\"odinger Bridge. We demonstrate the performance of\u0000our algorithm on a variety of unpaired data translation tasks.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269665","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Distributed Clustering based on Distributional Kernel 基于分布核的分布式聚类
arXiv - STAT - Machine Learning Pub Date : 2024-09-14 DOI: arxiv-2409.09418
Hang Zhang, Yang Xu, Lei Gong, Ye Zhu, Kai Ming Ting
{"title":"Distributed Clustering based on Distributional Kernel","authors":"Hang Zhang, Yang Xu, Lei Gong, Ye Zhu, Kai Ming Ting","doi":"arxiv-2409.09418","DOIUrl":"https://doi.org/arxiv-2409.09418","url":null,"abstract":"This paper introduces a new framework for clustering in a distributed network\u0000called Distributed Clustering based on Distributional Kernel (K) or KDC that\u0000produces the final clusters based on the similarity with respect to the\u0000distributions of initial clusters, as measured by K. It is the only framework\u0000that satisfies all three of the following properties. First, KDC guarantees\u0000that the combined clustering outcome from all sites is equivalent to the\u0000clustering outcome of its centralized counterpart from the combined dataset\u0000from all sites. Second, the maximum runtime cost of any site in distributed\u0000mode is smaller than the runtime cost in centralized mode. Third, it is\u0000designed to discover clusters of arbitrary shapes, sizes and densities. To the\u0000best of our knowledge, this is the first distributed clustering framework that\u0000employs a distributional kernel. The distribution-based clustering leads\u0000directly to significantly better clustering outcomes than existing methods of\u0000distributed clustering. In addition, we introduce a new clustering algorithm\u0000called Kernel Bounded Cluster Cores, which is the best clustering algorithm\u0000applied to KDC among existing clustering algorithms. We also show that KDC is a\u0000generic framework that enables a quadratic time clustering algorithm to deal\u0000with large datasets that would otherwise be impossible.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142269663","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Model-independent variable selection via the rule-based variable priorit 通过基于规则的变量优先级选择与模型无关的变量
arXiv - STAT - Machine Learning Pub Date : 2024-09-13 DOI: arxiv-2409.09003
Min Lu, Hemant Ishwaran
{"title":"Model-independent variable selection via the rule-based variable priorit","authors":"Min Lu, Hemant Ishwaran","doi":"arxiv-2409.09003","DOIUrl":"https://doi.org/arxiv-2409.09003","url":null,"abstract":"While achieving high prediction accuracy is a fundamental goal in machine\u0000learning, an equally important task is finding a small number of features with\u0000high explanatory power. One popular selection technique is permutation\u0000importance, which assesses a variable's impact by measuring the change in\u0000prediction error after permuting the variable. However, this can be problematic\u0000due to the need to create artificial data, a problem shared by other methods as\u0000well. Another problem is that variable selection methods can be limited by\u0000being model-specific. We introduce a new model-independent approach, Variable\u0000Priority (VarPro), which works by utilizing rules without the need to generate\u0000artificial data or evaluate prediction error. The method is relatively easy to\u0000use, requiring only the calculation of sample averages of simple statistics,\u0000and can be applied to many data settings, including regression, classification,\u0000and survival. We investigate the asymptotic properties of VarPro and show,\u0000among other things, that VarPro has a consistent filtering property for noise\u0000variables. Empirical studies using synthetic and real-world data show the\u0000method achieves a balanced performance and compares favorably to many\u0000state-of-the-art procedures currently used for variable selection.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261752","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Batched Online Contextual Sparse Bandits with Sequential Inclusion of Features 连续包含特征的分批在线上下文稀疏匪帮
arXiv - STAT - Machine Learning Pub Date : 2024-09-13 DOI: arxiv-2409.09199
Rowan Swiers, Subash Prabanantham, Andrew Maher
{"title":"Batched Online Contextual Sparse Bandits with Sequential Inclusion of Features","authors":"Rowan Swiers, Subash Prabanantham, Andrew Maher","doi":"arxiv-2409.09199","DOIUrl":"https://doi.org/arxiv-2409.09199","url":null,"abstract":"Multi-armed Bandits (MABs) are increasingly employed in online platforms and\u0000e-commerce to optimize decision making for personalized user experiences. In\u0000this work, we focus on the Contextual Bandit problem with linear rewards, under\u0000conditions of sparsity and batched data. We address the challenge of fairness\u0000by excluding irrelevant features from decision-making processes using a novel\u0000algorithm, Online Batched Sequential Inclusion (OBSI), which sequentially\u0000includes features as confidence in their impact on the reward increases. Our\u0000experiments on synthetic data show the superior performance of OBSI compared to\u0000other algorithms in terms of regret, relevance of features used, and compute.","PeriodicalId":501340,"journal":{"name":"arXiv - STAT - Machine Learning","volume":null,"pages":null},"PeriodicalIF":0.0,"publicationDate":"2024-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"142261751","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信