{"title":"LAV最佳子集过程中的惩罚计算和分支规则","authors":"R. Armstrong, P. Beck","doi":"10.1002/NAV.3800320305","DOIUrl":null,"url":null,"abstract":"Least absolute value (LAV) regression has become a widely accepted alternative to least squares regression. This has come about as the result of advancements in statistical theory and computational procedures to obtain LAV estimates. Computer codes are currently available to solve a wide range of LAV problems including the best subset regression. The purpose of this article is to study the use of penalty calculations and other branching rules in developing the solution tree for the best subset LAV regression.","PeriodicalId":431817,"journal":{"name":"Naval Research Logistics Quarterly","volume":"67 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1985-08-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Penalty calculations and branching rules in a LAV best subset procedure\",\"authors\":\"R. Armstrong, P. Beck\",\"doi\":\"10.1002/NAV.3800320305\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Least absolute value (LAV) regression has become a widely accepted alternative to least squares regression. This has come about as the result of advancements in statistical theory and computational procedures to obtain LAV estimates. Computer codes are currently available to solve a wide range of LAV problems including the best subset regression. The purpose of this article is to study the use of penalty calculations and other branching rules in developing the solution tree for the best subset LAV regression.\",\"PeriodicalId\":431817,\"journal\":{\"name\":\"Naval Research Logistics Quarterly\",\"volume\":\"67 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1985-08-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Naval Research Logistics Quarterly\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1002/NAV.3800320305\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Naval Research Logistics Quarterly","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1002/NAV.3800320305","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Penalty calculations and branching rules in a LAV best subset procedure
Least absolute value (LAV) regression has become a widely accepted alternative to least squares regression. This has come about as the result of advancements in statistical theory and computational procedures to obtain LAV estimates. Computer codes are currently available to solve a wide range of LAV problems including the best subset regression. The purpose of this article is to study the use of penalty calculations and other branching rules in developing the solution tree for the best subset LAV regression.