arXiv: Learning最新文献

筛选
英文 中文
Sampled Weighted Min-Hashing for Large-Scale Topic Mining 大规模主题挖掘的抽样加权最小哈希
arXiv: Learning Pub Date : 2015-06-24 DOI: 10.1007/978-3-319-19264-2_20
Gibran Fuentes-Pineda, Ivan Vladimir Meza Ruiz
{"title":"Sampled Weighted Min-Hashing for Large-Scale Topic Mining","authors":"Gibran Fuentes-Pineda, Ivan Vladimir Meza Ruiz","doi":"10.1007/978-3-319-19264-2_20","DOIUrl":"https://doi.org/10.1007/978-3-319-19264-2_20","url":null,"abstract":"","PeriodicalId":8468,"journal":{"name":"arXiv: Learning","volume":"59 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2015-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"87121769","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 1
Elicitation complexity of statistical properties 统计性质的引出复杂性
arXiv: Learning Pub Date : 2015-06-23 DOI: 10.1093/biomet/asaa093
Rafael M. Frongillo, Ian A. Kash
{"title":"Elicitation complexity of statistical properties","authors":"Rafael M. Frongillo, Ian A. Kash","doi":"10.1093/biomet/asaa093","DOIUrl":"https://doi.org/10.1093/biomet/asaa093","url":null,"abstract":"A property, or statistical functional, is said to be elicitable if it minimizes expected loss for some loss function. The study of which properties are elicitable sheds light on the capabilities and limitations of point estimation and empirical risk minimization. While recent work asks which properties are elicitable, we instead advocate for a more nuanced question: how many dimensions are required to indirectly elicit a given property? This number is called the elicitation complexity of the property. We lay the foundation for a general theory of elicitation complexity, including several basic results about how elicitation complexity behaves, and the complexity of standard properties of interest. Building on this foundation, our main result gives tight complexity bounds for the broad class of Bayes risks. We apply these results to several properties of interest, including variance, entropy, norms, and several classes of financial risk measures. We conclude with discussion and open directions.","PeriodicalId":8468,"journal":{"name":"arXiv: Learning","volume":"2 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2015-06-23","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"89115875","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 28
Multi-View Factorization Machines 多视图分解机
arXiv: Learning Pub Date : 2015-06-03 DOI: 10.1145/2835776.2835777
Bokai Cao, Hucheng Zhou, Guoqiang Li, Philip S. Yu
{"title":"Multi-View Factorization Machines","authors":"Bokai Cao, Hucheng Zhou, Guoqiang Li, Philip S. Yu","doi":"10.1145/2835776.2835777","DOIUrl":"https://doi.org/10.1145/2835776.2835777","url":null,"abstract":"For a learning task, data can usually be collected from different sources or be represented from multiple views. For example, laboratory results from different medical examinations are available for disease diagnosis, and each of them can only reflect the health state of a person from a particular aspect/view. Therefore, different views provide complementary information for learning tasks. An effective integration of the multi-view information is expected to facilitate the learning performance. In this paper, we propose a general predictor, named multi-view machines (MVMs), that can effectively include all the possible interactions between features from multiple views. A joint factorization is embedded for the full-order interaction parameters which allows parameter estimation under sparsity. Moreover, MVMs can work in conjunction with different loss functions for a variety of machine learning tasks. A stochastic gradient descent method is presented to learn the MVM model. We further illustrate the advantages of MVMs through comparison with other methods for multi-view classification, including support vector machines (SVMs), support tensor machines (STMs) and factorization machines (FMs).","PeriodicalId":8468,"journal":{"name":"arXiv: Learning","volume":"757 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2015-06-03","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"78802882","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 31
Proficiency Comparison of LADTree and REPTree Classifiers for Credit Risk Forecast LADTree和REPTree分类器在信用风险预测中的熟练程度比较
arXiv: Learning Pub Date : 2015-02-28 DOI: 10.5121/IJCSA.2015.5104
L. Devasena
{"title":"Proficiency Comparison of LADTree and REPTree Classifiers for Credit Risk Forecast","authors":"L. Devasena","doi":"10.5121/IJCSA.2015.5104","DOIUrl":"https://doi.org/10.5121/IJCSA.2015.5104","url":null,"abstract":"Predicting the Credit Defaulter is a perilous task of Financial Industries like Banks. Ascertaining non-payer before giving loan is a significant and conflict-ridden task of the Banker. Classification techniques are the better choice for predictive analysis like finding the claimant, whether he/she is an unpretentious customer or a cheat. Defining the outstanding classifier is a risky assignment for any industrialist like a banker. This allow computer science researchers to drill down efficient research works through evaluating different classifiers and finding out the best classifier for such predictive problems. This research work investigates the productivity of LADTree Classifier and REPTree Classifier for the credit risk prediction and compares their fitness through various measures. German credit dataset has been taken and used to predict the credit risk with a help of open source machine learning tool.","PeriodicalId":8468,"journal":{"name":"arXiv: Learning","volume":"4 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2015-02-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"74370408","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 4
Fast, Robust and Non-convex Subspace Recovery 快速、鲁棒和非凸子空间恢复
arXiv: Learning Pub Date : 2014-06-24 DOI: 10.1093/imaiai/iax012
Gilad Lerman, Tyler Maunu
{"title":"Fast, Robust and Non-convex Subspace Recovery","authors":"Gilad Lerman, Tyler Maunu","doi":"10.1093/imaiai/iax012","DOIUrl":"https://doi.org/10.1093/imaiai/iax012","url":null,"abstract":"This work presents a fast and non-convex algorithm for robust subspace recovery. The data sets considered include inliers drawn around a low-dimensional subspace of a higher dimensional ambient space, and a possibly large portion of outliers that do not lie nearby this subspace. The proposed algorithm, which we refer to as Fast Median Subspace (FMS), is designed to robustly determine the underlying subspace of such data sets, while having lower computational complexity than existing methods. We prove convergence of the FMS iterates to a stationary point. Further, under a special model of data, FMS converges to a point which is near to the global minimum with overwhelming probability. Under this model, we show that the iteration complexity is globally bounded and locally $r$-linear. The latter theorem holds for any fixed fraction of outliers (less than 1) and any fixed positive distance between the limit point and the global minimum. Numerical experiments on synthetic and real data demonstrate its competitive speed and accuracy.","PeriodicalId":8468,"journal":{"name":"arXiv: Learning","volume":"96 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2014-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"79068151","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 67
Cross-situational and supervised learning in the emergence of communication 交际出现中的跨情境和监督学习
arXiv: Learning Pub Date : 2009-01-26 DOI: 10.1075/is.12.1.05fon
J. Fontanari, A. Cangelosi
{"title":"Cross-situational and supervised learning in the emergence of communication","authors":"J. Fontanari, A. Cangelosi","doi":"10.1075/is.12.1.05fon","DOIUrl":"https://doi.org/10.1075/is.12.1.05fon","url":null,"abstract":"Scenarios for the emergence or bootstrap of a lexicon involve the repeated interaction between at least two agents who must reach a consensus on how to name N objects using H words. Here we consider minimal models of two types of learning algorithms: cross-situational learning, in which the individuals determine the meaning of a word by looking for something in common across all observed uses of that word, and supervised operant conditioning learning, in which there is strong feedback between individuals about the intended meaning of the words. Despite the stark differences between these learning schemes, we show that they yield the same communication accuracy in the realistic limits of large N and H, which coincides with the result of the classical occupancy problem of randomly assigning N objects to H words.","PeriodicalId":8468,"journal":{"name":"arXiv: Learning","volume":"7 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2009-01-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"85366010","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 10
Entropy, Perception, and Relativity 熵、知觉和相对性
arXiv: Learning Pub Date : 2006-04-01 DOI: 10.1037/e645982007-001
S. Jaegar
{"title":"Entropy, Perception, and Relativity","authors":"S. Jaegar","doi":"10.1037/e645982007-001","DOIUrl":"https://doi.org/10.1037/e645982007-001","url":null,"abstract":"Abstract : In this paper, I expand Shannon's definition of entropy into a new form of entropy that allows integration of information from different random events. Shannon's notion of entropy is a special case of my more general definition of entropy. I define probability using a so-called performance function, which is de facto an exponential distribution. Assuming that my general notion of entropy reflects the true uncertainty about a probabilistic event, I understand that our perceived uncertainty differs. I claim that our perception is the result of two opposing forces similar to the two famous antagonists in Chinese philosophy: Yin and Yang. Based on this idea, I show that our perceived uncertainty matches the true uncertainty in points determined by the golden ratio. I demonstrate that the well-known sigmoid function, which we typically employ in artificial neural networks as a non-linear threshold function, describes the actual performance. Furthermore, I provide a motivation for the time dilation in Einstein's Special Relativity, basically claiming that although time dilation conforms with our perception, it does not correspond to reality. At the end of the paper, I show how to apply this theoretical framework to practical applications. I present recognition rates for a pattern recognition problem, and also propose a network architecture that can take advantage of general entropy to solve complex decision problems.","PeriodicalId":8468,"journal":{"name":"arXiv: Learning","volume":"53 1","pages":""},"PeriodicalIF":0.0,"publicationDate":"2006-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"83051930","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信