{"title":"直接使用共轭梯度的逻辑回归的有效优化","authors":"Kenji Watanabe, Takumi Kobayashi, N. Otsu","doi":"10.1109/ICMLA.2011.63","DOIUrl":null,"url":null,"abstract":"In classification problems, logistic regression (LR) is used to estimate posterior probabilities. The objective function of LR is usually minimized by Newton-Raphson method such as using iterative reweighted least squares (IRLS). There, the inverse Hessian matrix must be calculated in each iteration step. Thus, a computational cost in the optimization of LR significantly increases as input data becomes large. To reduce the computational cost, we propose a novel optimization method of LR by directly using the non-linear conjugate gradient (CG) method. The proposed method iteratively minimizes the objective function of LR without calculation of the Hessian matrix. Furthermore, to reduce the number of iteration efficiently, the step size in the non-linear CG iteration is optimized avoiding ad hock line search, and initial values are set by ordinary linear regression analysis. In the experimental results, our method performs about 200 times faster than the other methods for a large scale dataset.","PeriodicalId":439926,"journal":{"name":"2011 10th International Conference on Machine Learning and Applications and Workshops","volume":"7 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2011-12-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Efficient Optimization of Logistic Regression by Direct Use of Conjugate Gradient\",\"authors\":\"Kenji Watanabe, Takumi Kobayashi, N. Otsu\",\"doi\":\"10.1109/ICMLA.2011.63\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"In classification problems, logistic regression (LR) is used to estimate posterior probabilities. The objective function of LR is usually minimized by Newton-Raphson method such as using iterative reweighted least squares (IRLS). There, the inverse Hessian matrix must be calculated in each iteration step. Thus, a computational cost in the optimization of LR significantly increases as input data becomes large. To reduce the computational cost, we propose a novel optimization method of LR by directly using the non-linear conjugate gradient (CG) method. The proposed method iteratively minimizes the objective function of LR without calculation of the Hessian matrix. Furthermore, to reduce the number of iteration efficiently, the step size in the non-linear CG iteration is optimized avoiding ad hock line search, and initial values are set by ordinary linear regression analysis. In the experimental results, our method performs about 200 times faster than the other methods for a large scale dataset.\",\"PeriodicalId\":439926,\"journal\":{\"name\":\"2011 10th International Conference on Machine Learning and Applications and Workshops\",\"volume\":\"7 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2011-12-18\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2011 10th International Conference on Machine Learning and Applications and Workshops\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICMLA.2011.63\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2011 10th International Conference on Machine Learning and Applications and Workshops","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICMLA.2011.63","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Efficient Optimization of Logistic Regression by Direct Use of Conjugate Gradient
In classification problems, logistic regression (LR) is used to estimate posterior probabilities. The objective function of LR is usually minimized by Newton-Raphson method such as using iterative reweighted least squares (IRLS). There, the inverse Hessian matrix must be calculated in each iteration step. Thus, a computational cost in the optimization of LR significantly increases as input data becomes large. To reduce the computational cost, we propose a novel optimization method of LR by directly using the non-linear conjugate gradient (CG) method. The proposed method iteratively minimizes the objective function of LR without calculation of the Hessian matrix. Furthermore, to reduce the number of iteration efficiently, the step size in the non-linear CG iteration is optimized avoiding ad hock line search, and initial values are set by ordinary linear regression analysis. In the experimental results, our method performs about 200 times faster than the other methods for a large scale dataset.