{"title":"正交约束优化的改进最陡下降和牛顿算法。第二部分。复数格拉斯曼流形","authors":"J. Manton","doi":"10.1109/ISSPA.2001.949781","DOIUrl":null,"url":null,"abstract":"The classical steepest descent and Newton algorithms can be used to minimise a cost function f(X). This paper shows how they can be modified to take into account the constraint that the columns of the complex-valued matrix X are mutually orthogonal and have unit norm. It is assumed that the cost function satisfies f(XQ) = f(X) for any unitary matrix Q. This allows the constrained optimisation problem to be converted into an unconstrained one on the Grassmann manifold. This significantly reduces the dimension of the optimisation problem and often results in faster convergence.","PeriodicalId":236050,"journal":{"name":"Proceedings of the Sixth International Symposium on Signal Processing and its Applications (Cat.No.01EX467)","volume":"46 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"7","resultStr":"{\"title\":\"Modified steepest descent and Newton algorithms for orthogonally constrained optimisation. Part II. The complex Grassmann manifold\",\"authors\":\"J. Manton\",\"doi\":\"10.1109/ISSPA.2001.949781\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The classical steepest descent and Newton algorithms can be used to minimise a cost function f(X). This paper shows how they can be modified to take into account the constraint that the columns of the complex-valued matrix X are mutually orthogonal and have unit norm. It is assumed that the cost function satisfies f(XQ) = f(X) for any unitary matrix Q. This allows the constrained optimisation problem to be converted into an unconstrained one on the Grassmann manifold. This significantly reduces the dimension of the optimisation problem and often results in faster convergence.\",\"PeriodicalId\":236050,\"journal\":{\"name\":\"Proceedings of the Sixth International Symposium on Signal Processing and its Applications (Cat.No.01EX467)\",\"volume\":\"46 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1900-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"7\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the Sixth International Symposium on Signal Processing and its Applications (Cat.No.01EX467)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ISSPA.2001.949781\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the Sixth International Symposium on Signal Processing and its Applications (Cat.No.01EX467)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ISSPA.2001.949781","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Modified steepest descent and Newton algorithms for orthogonally constrained optimisation. Part II. The complex Grassmann manifold
The classical steepest descent and Newton algorithms can be used to minimise a cost function f(X). This paper shows how they can be modified to take into account the constraint that the columns of the complex-valued matrix X are mutually orthogonal and have unit norm. It is assumed that the cost function satisfies f(XQ) = f(X) for any unitary matrix Q. This allows the constrained optimisation problem to be converted into an unconstrained one on the Grassmann manifold. This significantly reduces the dimension of the optimisation problem and often results in faster convergence.