Principal Component Analysis

Xuan Chen
{"title":"Principal Component Analysis","authors":"Xuan Chen","doi":"10.1142/9781786349378_0007","DOIUrl":null,"url":null,"abstract":"Xuanye Chen Introduction Principal component analysis was first introduced by Karl Pearson for non-random variables, and then H. Hotelling extended this method to the case of random vectors. Principal component analysis (PCA) is a technique for reducing dimensionality, increasing interpretability, and at the same time minimizing information loss. Definition Principal Component Analysis (PCA) is a statistical method. Through orthogonal transformation, a group of variables that may be correlated is transformed into a group of linearly uncorrelated variables, which are called principal components. To be specific, it transforms the data to a new coordinate system such that the greatest variance by some scalar projection of the data comes to lie on the first coordinate (called the first principal component), the second greatest variance on the second coordinate, and so on. The Calculation of PCA F1 is used to represent the first linear combination selected, that is, the first comprehensive indicator. The larger the Var ( F1 ) is, the more information F1 contains. Therefore, F1 selected among all linear combinations has the largest variance, so F1 is called the first principal component. If the first principal component is not enough to represent the information of the original P indicators, then F2 is selected, that is, the second linear combination. In order to effectively reflect the original information, the existing information of F1 does not need to appear in F2. In other words, Cov(F1, F2) = 0, and F2 is called the second principal component. And so on, we can construct 3rd, 4th, ... , Pth principal component. Fp = a1i*ZX1 + a2i*ZX2 + ...... + api*ZXp","PeriodicalId":402819,"journal":{"name":"Advanced Textbooks in Mathematics","volume":"59 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advanced Textbooks in Mathematics","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1142/9781786349378_0007","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Xuanye Chen Introduction Principal component analysis was first introduced by Karl Pearson for non-random variables, and then H. Hotelling extended this method to the case of random vectors. Principal component analysis (PCA) is a technique for reducing dimensionality, increasing interpretability, and at the same time minimizing information loss. Definition Principal Component Analysis (PCA) is a statistical method. Through orthogonal transformation, a group of variables that may be correlated is transformed into a group of linearly uncorrelated variables, which are called principal components. To be specific, it transforms the data to a new coordinate system such that the greatest variance by some scalar projection of the data comes to lie on the first coordinate (called the first principal component), the second greatest variance on the second coordinate, and so on. The Calculation of PCA F1 is used to represent the first linear combination selected, that is, the first comprehensive indicator. The larger the Var ( F1 ) is, the more information F1 contains. Therefore, F1 selected among all linear combinations has the largest variance, so F1 is called the first principal component. If the first principal component is not enough to represent the information of the original P indicators, then F2 is selected, that is, the second linear combination. In order to effectively reflect the original information, the existing information of F1 does not need to appear in F2. In other words, Cov(F1, F2) = 0, and F2 is called the second principal component. And so on, we can construct 3rd, 4th, ... , Pth principal component. Fp = a1i*ZX1 + a2i*ZX2 + ...... + api*ZXp
主成分分析
主成分分析首先由Karl Pearson引入到非随机变量中,随后H. Hotelling将该方法推广到随机向量的情况。主成分分析(PCA)是一种减少维数、增加可解释性、同时最小化信息损失的技术。主成分分析(PCA)是一种统计方法。通过正交变换,将一组可能相关的变量转化为一组线性不相关的变量,称为主成分。具体地说,它将数据转换为一个新的坐标系,使得数据的某个标量投影的最大方差位于第一个坐标上(称为第一个主成分),第二个最大方差位于第二个坐标上,以此类推。用PCA的计算F1表示选择的第一个线性组合,即第一个综合指标。Var (F1)越大,F1包含的信息越多。因此,在所有线性组合中选择的F1方差最大,因此称F1为第一主成分。如果第一个主成分不足以表示原P个指标的信息,则选择F2,即第二次线性组合。为了有效地反映原始信息,F1的现有信息不需要出现在F2中。也就是说,Cov(F1, F2) = 0, F2称为第二主成分。以此类推,我们可以构造3、4、…第p个主成分。Fp = a1i*ZX1 + a2i*ZX2 + ......+ api * ZXp
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信