Fundamentals of Machine Learning最新文献

筛选
英文 中文
Reinforcement learning 强化学习
Fundamentals of Machine Learning Pub Date : 2019-11-28 DOI: 10.1093/oso/9780198828044.003.0010
T. Trappenberg
{"title":"Reinforcement learning","authors":"T. Trappenberg","doi":"10.1093/oso/9780198828044.003.0010","DOIUrl":"https://doi.org/10.1093/oso/9780198828044.003.0010","url":null,"abstract":"The discussion here considers a much more common learning condition where an agent, such as a human or a robot, has to learn to make decisions in the environment from simple feedback. Such feedback is provided only after periods of actions in the form of reward or punishment without detailing which of the actions has contributed to the outcome. This type of learning scenario is called reinforcement learning. This learning problem is formalized in a Markov decision-making process with a variety of related algorithms. The second part of this chapter will use function approximators with neural networks which have made recent progress as deep reinforcement learning.","PeriodicalId":300420,"journal":{"name":"Fundamentals of Machine Learning","volume":"23 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"133710211","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Generative models 生成模型
Fundamentals of Machine Learning Pub Date : 2019-11-28 DOI: 10.1093/oso/9780198828044.003.0008
T. Trappenberg
{"title":"Generative models","authors":"T. Trappenberg","doi":"10.1093/oso/9780198828044.003.0008","DOIUrl":"https://doi.org/10.1093/oso/9780198828044.003.0008","url":null,"abstract":"This chapter presents an introduction to the important topic of building generative models. These are models that are aimed to understand the variety of a class such as cars or trees. A generative mode should be able to generate feature vectors for instances of the class they represent, and such models should therefore be able to characterize the class with all its variations. The subject is discussed both in a Bayesian and in a deep learning context, and also within a supervised and unsupervised context. This area is related to important algorithms such as k-means clustering, expectation maximization (EM), naïve Bayes, generative adversarial networks (GANs), and variational autoencoders (VAE).","PeriodicalId":300420,"journal":{"name":"Fundamentals of Machine Learning","volume":"39 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"127900018","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
Basic probability theory 基本概率论
Fundamentals of Machine Learning Pub Date : 2019-11-28 DOI: 10.1093/oso/9780198828044.003.0006
T. Trappenberg
{"title":"Basic probability theory","authors":"T. Trappenberg","doi":"10.1093/oso/9780198828044.003.0006","DOIUrl":"https://doi.org/10.1093/oso/9780198828044.003.0006","url":null,"abstract":"The discussion provides a refresher of probability theory, in particular with respect to the formulations that build the theoretical language of modern machine learning. Probability theory is the formalism of random numbers, and this chapter outlines what these are and how they are characterized by probability density or probability mass functions. How such functions have traditionally been characterized is covered, and a review of how to work with such mathematical objects such as transforming density functions and how to measure differences between density function is presented. Definitions and basic operations with multiple random variables, including the Bayes law, are covered. The chapter ends with an outline of some important approximation techniques of so-called Monte Carlo methods.","PeriodicalId":300420,"journal":{"name":"Fundamentals of Machine Learning","volume":"1 1","pages":"0"},"PeriodicalIF":0.0,"publicationDate":"2019-11-28","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"128918179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":0,"RegionCategory":"","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
引用次数: 0
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
相关产品
×
本文献相关产品
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信