{"title":"时空自适应处理算法的高效训练策略","authors":"G.K. Borsari, A. Steinhardt","doi":"10.1109/ACSSC.1995.540629","DOIUrl":null,"url":null,"abstract":"Space-time adaptive processing (STAP) usually requires the estimation of large-dimension clutter covariance matrices. The mean loss in output SNR is a function of the number of statistically similar data samples used to estimate the covariance matrix. This number is generally 3 times the dimension of the covariance matrix or more. In nonhomogeneous clutter environments it is difficult to obtain this many statistically similar data samples using a data selection rule that is computationally simple. We present several new training strategies that select data samples from as close to the target range-gate as possible and simultaneously maintain a low computation count. A \"training strategy\" is the rule used to select data samples for covariance matrix estimation. A new training strategy is presented along with a recursion for efficient estimation of the clutter covariance matrix at each target range-gate. Also, a new training concept called freeze training is presented and shown to reduce the number of computations and to mitigate clutter discretes in nulled output data. A computation-count comparison is presented with each training strategy.","PeriodicalId":171264,"journal":{"name":"Conference Record of The Twenty-Ninth Asilomar Conference on Signals, Systems and Computers","volume":"103 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1995-10-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"18","resultStr":"{\"title\":\"Cost-efficient training strategies for space-time adaptive processing algorithms\",\"authors\":\"G.K. Borsari, A. Steinhardt\",\"doi\":\"10.1109/ACSSC.1995.540629\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Space-time adaptive processing (STAP) usually requires the estimation of large-dimension clutter covariance matrices. The mean loss in output SNR is a function of the number of statistically similar data samples used to estimate the covariance matrix. This number is generally 3 times the dimension of the covariance matrix or more. In nonhomogeneous clutter environments it is difficult to obtain this many statistically similar data samples using a data selection rule that is computationally simple. We present several new training strategies that select data samples from as close to the target range-gate as possible and simultaneously maintain a low computation count. A \\\"training strategy\\\" is the rule used to select data samples for covariance matrix estimation. A new training strategy is presented along with a recursion for efficient estimation of the clutter covariance matrix at each target range-gate. Also, a new training concept called freeze training is presented and shown to reduce the number of computations and to mitigate clutter discretes in nulled output data. A computation-count comparison is presented with each training strategy.\",\"PeriodicalId\":171264,\"journal\":{\"name\":\"Conference Record of The Twenty-Ninth Asilomar Conference on Signals, Systems and Computers\",\"volume\":\"103 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1995-10-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"18\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Conference Record of The Twenty-Ninth Asilomar Conference on Signals, Systems and Computers\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ACSSC.1995.540629\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Conference Record of The Twenty-Ninth Asilomar Conference on Signals, Systems and Computers","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ACSSC.1995.540629","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Cost-efficient training strategies for space-time adaptive processing algorithms
Space-time adaptive processing (STAP) usually requires the estimation of large-dimension clutter covariance matrices. The mean loss in output SNR is a function of the number of statistically similar data samples used to estimate the covariance matrix. This number is generally 3 times the dimension of the covariance matrix or more. In nonhomogeneous clutter environments it is difficult to obtain this many statistically similar data samples using a data selection rule that is computationally simple. We present several new training strategies that select data samples from as close to the target range-gate as possible and simultaneously maintain a low computation count. A "training strategy" is the rule used to select data samples for covariance matrix estimation. A new training strategy is presented along with a recursion for efficient estimation of the clutter covariance matrix at each target range-gate. Also, a new training concept called freeze training is presented and shown to reduce the number of computations and to mitigate clutter discretes in nulled output data. A computation-count comparison is presented with each training strategy.