International Conference on Machine Learning, Optimization, and Data Science最新文献

筛选
英文 中文
A Comparison of SVM against Pre-trained Language Models (PLMs) for Text Classification Tasks 支持向量机与预训练语言模型在文本分类任务中的比较
International Conference on Machine Learning, Optimization, and Data Science Pub Date : 2022-11-04 DOI: 10.48550/arXiv.2211.02563
Yasmen Wahba, N. Madhavji, John Steinbacher
{"title":"A Comparison of SVM against Pre-trained Language Models (PLMs) for Text Classification Tasks","authors":"Yasmen Wahba, N. Madhavji, John Steinbacher","doi":"10.48550/arXiv.2211.02563","DOIUrl":"https://doi.org/10.48550/arXiv.2211.02563","url":null,"abstract":"The emergence of pre-trained language models (PLMs) has shown great success in many Natural Language Processing (NLP) tasks including text classification. Due to the minimal to no feature engineering required when using these models, PLMs are becoming the de facto choice for any NLP task. However, for domain-specific corpora (e.g., financial, legal, and industrial), fine-tuning a pre-trained model for a specific task has shown to provide a performance improvement. In this paper, we compare the performance of four different PLMs on three public domain-free datasets and a real-world dataset containing domain-specific words, against a simple SVM linear classifier with TFIDF vectorized text. The experimental results on the four datasets show that using PLMs, even fine-tuned, do not provide significant gain over the linear SVM classifier. Hence, we recommend that for text classification tasks, traditional SVM along with careful feature engineering can pro-vide a cheaper and superior performance than PLMs.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"29 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-11-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130027693","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 7
Hierarchical Decentralized Deep Reinforcement Learning Architecture for a Simulated Four-Legged Agent 模拟四足智能体的分层分散深度强化学习体系结构
International Conference on Machine Learning, Optimization, and Data Science Pub Date : 2022-09-21 DOI: 10.48550/arXiv.2210.08003
W. Z. E. Amri, L. Hermes, M. Schilling
{"title":"Hierarchical Decentralized Deep Reinforcement Learning Architecture for a Simulated Four-Legged Agent","authors":"W. Z. E. Amri, L. Hermes, M. Schilling","doi":"10.48550/arXiv.2210.08003","DOIUrl":"https://doi.org/10.48550/arXiv.2210.08003","url":null,"abstract":"Legged locomotion is widespread in nature and has inspired the design of current robots. The controller of these legged robots is often realized as one centralized instance. However, in nature, control of movement happens in a hierarchical and decentralized fashion. Introducing these biological design principles into robotic control systems has motivated this work. We tackle the question whether decentralized and hierarchical control is beneficial for legged robots and present a novel decentral, hierarchical architecture to control a simulated legged agent. Three different tasks varying in complexity are designed to benchmark five architectures (centralized, decentralized, hierarchical and two different combinations of hierarchical decentralized architectures). The results demonstrate that decentralizing the different levels of the hierarchical architectures facilitates learning of the agent, ensures more energy efficient movements as well as robustness towards new unseen environments. Furthermore, this comparison sheds light on the importance of modularity in hierarchical architectures to solve complex goal-directed tasks. We provide an open-source code implementation of our architecture (https://github.com/wzaielamri/hddrl).","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"481 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"114782452","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
On the utility and protection of optimization with differential privacy and classic regularization techniques 差分隐私和经典正则化技术优化的效用和保护
International Conference on Machine Learning, Optimization, and Data Science Pub Date : 2022-09-07 DOI: 10.48550/arXiv.2209.03175
Eugenio Lomurno, Matteo Matteucci
{"title":"On the utility and protection of optimization with differential privacy and classic regularization techniques","authors":"Eugenio Lomurno, Matteo Matteucci","doi":"10.48550/arXiv.2209.03175","DOIUrl":"https://doi.org/10.48550/arXiv.2209.03175","url":null,"abstract":"Nowadays, owners and developers of deep learning models must consider stringent privacy-preservation rules of their training data, usually crowd-sourced and retaining sensitive information. The most widely adopted method to enforce privacy guarantees of a deep learning model nowadays relies on optimization techniques enforcing differential privacy. According to the literature, this approach has proven to be a successful defence against several models' privacy attacks, but its downside is a substantial degradation of the models' performance. In this work, we compare the effectiveness of the differentially-private stochastic gradient descent (DP-SGD) algorithm against standard optimization practices with regularization techniques. We analyze the resulting models' utility, training performance, and the effectiveness of membership inference and model inversion attacks against the learned models. Finally, we discuss differential privacy's flaws and limits and empirically demonstrate the often superior privacy-preserving properties of dropout and l2-regularization.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"40 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-09-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130289948","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 2
Adaptive Zeroth-Order Optimisation of Nonconvex Composite Objectives 非凸复合目标的自适应零阶优化
International Conference on Machine Learning, Optimization, and Data Science Pub Date : 2022-08-09 DOI: 10.48550/arXiv.2208.04579
Weijia Shao, S. Albayrak
{"title":"Adaptive Zeroth-Order Optimisation of Nonconvex Composite Objectives","authors":"Weijia Shao, S. Albayrak","doi":"10.48550/arXiv.2208.04579","DOIUrl":"https://doi.org/10.48550/arXiv.2208.04579","url":null,"abstract":"In this paper, we propose and analyze algorithms for zeroth-order optimization of non-convex composite objectives, focusing on reducing the complexity dependence on dimensionality. This is achieved by exploiting the low dimensional structure of the decision set using the stochastic mirror descent method with an entropy alike function, which performs gradient descent in the space equipped with the maximum norm. To improve the gradient estimation, we replace the classic Gaussian smoothing method with a sampling method based on the Rademacher distribution and show that the mini-batch method copes with the non-Euclidean geometry. To avoid tuning hyperparameters, we analyze the adaptive stepsizes for the general stochastic mirror descent and show that the adaptive version of the proposed algorithm converges without requiring prior knowledge about the problem.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"13 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-09","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"132546629","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Robust PCA for Anomaly Detection and Data Imputation in Seasonal Time Series 基于鲁棒主成分分析的季节时间序列异常检测与数据代入
International Conference on Machine Learning, Optimization, and Data Science Pub Date : 2022-08-03 DOI: 10.48550/arXiv.2208.01998
Hông-Lan Botterman, Julien Roussel, Thomas Morzadec, A. Jabbari, N. Brunel
{"title":"Robust PCA for Anomaly Detection and Data Imputation in Seasonal Time Series","authors":"Hông-Lan Botterman, Julien Roussel, Thomas Morzadec, A. Jabbari, N. Brunel","doi":"10.48550/arXiv.2208.01998","DOIUrl":"https://doi.org/10.48550/arXiv.2208.01998","url":null,"abstract":"We propose a robust principal component analysis (RPCA) framework to recover low-rank and sparse matrices from temporal observations. We develop an online version of the batch temporal algorithm in order to process larger datasets or streaming data. We empirically compare the proposed approaches with different RPCA frameworks and show their effectiveness in practical situations.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"4 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-08-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"134499057","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Parallel Bayesian Optimization of Agent-based Transportation Simulation 基于agent的交通仿真并行贝叶斯优化
International Conference on Machine Learning, Optimization, and Data Science Pub Date : 2022-07-11 DOI: 10.48550/arXiv.2207.05041
K. Chhatre, Sidney A. Feygin, C. Sheppard, R. Waraich
{"title":"Parallel Bayesian Optimization of Agent-based Transportation Simulation","authors":"K. Chhatre, Sidney A. Feygin, C. Sheppard, R. Waraich","doi":"10.48550/arXiv.2207.05041","DOIUrl":"https://doi.org/10.48550/arXiv.2207.05041","url":null,"abstract":"MATSim (Multi-Agent Transport Simulation Toolkit) is an open source large-scale agent-based transportation planning project applied to various areas like road transport, public transport, freight transport, regional evacuation, etc. BEAM (Behavior, Energy, Autonomy, and Mobility) framework extends MATSim to enable powerful and scalable analysis of urban transportation systems. The agents from the BEAM simulation exhibit 'mode choice' behavior based on multinomial logit model. In our study, we consider eight mode choices viz. bike, car, walk, ride hail, driving to transit, walking to transit, ride hail to transit, and ride hail pooling. The 'alternative specific constants' for each mode choice are critical hyperparameters in a configuration file related to a particular scenario under experimentation. We use the 'Urbansim-10k' BEAM scenario (with 10,000 population size) for all our experiments. Since these hyperparameters affect the simulation in complex ways, manual calibration methods are time consuming. We present a parallel Bayesian optimization method with early stopping rule to achieve fast convergence for the given multi-in-multi-out problem to its optimal configurations. Our model is based on an open source HpBandSter package. This approach combines hierarchy of several 1D Kernel Density Estimators (KDE) with a cheap evaluator (Hyperband, a single multidimensional KDE). Our model has also incorporated extrapolation based early stopping rule. With our model, we could achieve a 25% L1 norm for a large-scale BEAM simulation in fully autonomous manner. To the best of our knowledge, our work is the first of its kind applied to large-scale multi-agent transportation simulations. This work can be useful for surrogate modeling of scenarios with very large populations.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"61 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-07-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121080367","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Brain-like combination of feedforward and recurrent network components achieves prototype extraction and robust pattern recognition 将前馈和递归网络成分结合,实现了原型提取和鲁棒模式识别
International Conference on Machine Learning, Optimization, and Data Science Pub Date : 2022-06-30 DOI: 10.48550/arXiv.2206.15036
Naresh B. Ravichandran, A. Lansner, P. Herman
{"title":"Brain-like combination of feedforward and recurrent network components achieves prototype extraction and robust pattern recognition","authors":"Naresh B. Ravichandran, A. Lansner, P. Herman","doi":"10.48550/arXiv.2206.15036","DOIUrl":"https://doi.org/10.48550/arXiv.2206.15036","url":null,"abstract":"Associative memory has been a prominent candidate for the computation performed by the massively recurrent neocortical networks. Attractor networks implementing associative memory have offered mechanistic explanation for many cognitive phenomena. However, attractor memory models are typically trained using orthogonal or random patterns to avoid interference between memories, which makes them unfeasible for naturally occurring complex correlated stimuli like images. We approach this problem by combining a recurrent attractor network with a feedforward network that learns distributed representations using an unsupervised Hebbian-Bayesian learning rule. The resulting network model incorporates many known biological properties: unsupervised learning, Hebbian plasticity, sparse distributed activations, sparse connectivity, columnar and laminar cortical architecture, etc. We evaluate the synergistic effects of the feedforward and recurrent network components in complex pattern recognition tasks on the MNIST handwritten digits dataset. We demonstrate that the recurrent attractor component implements associative memory when trained on the feedforward-driven internal (hidden) representations. The associative memory is also shown to perform prototype extraction from the training data and make the representations robust to severely distorted input. We argue that several aspects of the proposed integration of feedforward and recurrent computations are particularly attractive from a machine learning perspective.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"9 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"130978720","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
MicroRacer: a didactic environment for Deep Reinforcement Learning MicroRacer:深度强化学习的教学环境
International Conference on Machine Learning, Optimization, and Data Science Pub Date : 2022-03-20 DOI: 10.48550/arXiv.2203.10494
A. Asperti, Marco Del Brutto
{"title":"MicroRacer: a didactic environment for Deep Reinforcement Learning","authors":"A. Asperti, Marco Del Brutto","doi":"10.48550/arXiv.2203.10494","DOIUrl":"https://doi.org/10.48550/arXiv.2203.10494","url":null,"abstract":"MicroRacer is a simple, open source environment inspired by car racing especially meant for the didactics of Deep Reinforcement Learning. The complexity of the environment has been explicitly calibrated to allow users to experiment with many different methods, networks and hyperparameters settings without requiring sophisticated software or the need of exceedingly long training times. Baseline agents for major learning algorithms such as DDPG, PPO, SAC, TD2 and DSAC are provided too, along with a preliminary comparison in terms of training time and performance.","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-03-20","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"121095916","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Algorithms that Get Old: The Case of Generative Deep Neural Networks 过时的算法:生成式深度神经网络的案例
International Conference on Machine Learning, Optimization, and Data Science Pub Date : 2022-02-07 DOI: 10.1007/978-3-031-25891-6_14
G. Turinici
{"title":"Algorithms that Get Old: The Case of Generative Deep Neural Networks","authors":"G. Turinici","doi":"10.1007/978-3-031-25891-6_14","DOIUrl":"https://doi.org/10.1007/978-3-031-25891-6_14","url":null,"abstract":"","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"18 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-02-07","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"126706235","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Brain Structural Saliency Over The Ages 多年来大脑结构的显著性
International Conference on Machine Learning, Optimization, and Data Science Pub Date : 2022-01-12 DOI: 10.1007/978-3-031-25891-6_40
Daniel Taylor, Jonathan Shock, Deshendran Moodley, J. Ipser, M. Treder
{"title":"Brain Structural Saliency Over The Ages","authors":"Daniel Taylor, Jonathan Shock, Deshendran Moodley, J. Ipser, M. Treder","doi":"10.1007/978-3-031-25891-6_40","DOIUrl":"https://doi.org/10.1007/978-3-031-25891-6_40","url":null,"abstract":"","PeriodicalId":432112,"journal":{"name":"International Conference on Machine Learning, Optimization, and Data Science","volume":"44 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2022-01-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133554114","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信