{"title":"One-Bit Quantization and Sparsification for Multiclass Linear Classification With Strong Regularization","authors":"Reza Ghane;Danil Akhtiamov;Babak Hassibi","doi":"10.1109/TSP.2025.3598246","DOIUrl":null,"url":null,"abstract":"We study the use of linear regression for multiclass classification in the over-parametrized regime where some of the training data is mislabeled. In such scenarios it is necessary to add an explicit regularization term, <inline-formula><tex-math>$\\lambda f(\\cdot)$</tex-math></inline-formula>, for some convex function <inline-formula><tex-math>$f(\\cdot)$</tex-math></inline-formula>, to avoid overfitting the mislabeled data. In our analysis, we assume that the data is sampled from a Gaussian Mixture Model with equal class sizes, and that a proportion of the training labels is corrupted for each class. Under these assumptions, we prove that the best classification performance is achieved when <inline-formula><tex-math>$f(\\cdot)=\\|\\cdot\\|^{2}_{2}$</tex-math></inline-formula> and <inline-formula><tex-math>$\\lambda\\to\\infty$</tex-math></inline-formula>. We then proceed to analyze the classification errors for <inline-formula><tex-math>$f(\\cdot)=\\|\\cdot\\|_{1}$</tex-math></inline-formula> and <inline-formula><tex-math>$f(\\cdot)=\\|\\cdot\\|_{\\infty}$</tex-math></inline-formula> in the large <inline-formula><tex-math>$\\lambda$</tex-math></inline-formula> regime and notice that it is often possible to find sparse and one-bit solutions, respectively, that perform almost as well as the one corresponding to <inline-formula><tex-math>$f(\\cdot)=\\|\\cdot\\|_{2}^{2}$</tex-math></inline-formula>.","PeriodicalId":13330,"journal":{"name":"IEEE Transactions on Signal Processing","volume":"73 ","pages":"3270-3285"},"PeriodicalIF":5.8000,"publicationDate":"2025-08-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"IEEE Transactions on Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://ieeexplore.ieee.org/document/11123530/","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0
Abstract
We study the use of linear regression for multiclass classification in the over-parametrized regime where some of the training data is mislabeled. In such scenarios it is necessary to add an explicit regularization term, $\lambda f(\cdot)$, for some convex function $f(\cdot)$, to avoid overfitting the mislabeled data. In our analysis, we assume that the data is sampled from a Gaussian Mixture Model with equal class sizes, and that a proportion of the training labels is corrupted for each class. Under these assumptions, we prove that the best classification performance is achieved when $f(\cdot)=\|\cdot\|^{2}_{2}$ and $\lambda\to\infty$. We then proceed to analyze the classification errors for $f(\cdot)=\|\cdot\|_{1}$ and $f(\cdot)=\|\cdot\|_{\infty}$ in the large $\lambda$ regime and notice that it is often possible to find sparse and one-bit solutions, respectively, that perform almost as well as the one corresponding to $f(\cdot)=\|\cdot\|_{2}^{2}$.
期刊介绍:
The IEEE Transactions on Signal Processing covers novel theory, algorithms, performance analyses and applications of techniques for the processing, understanding, learning, retrieval, mining, and extraction of information from signals. The term “signal” includes, among others, audio, video, speech, image, communication, geophysical, sonar, radar, medical and musical signals. Examples of topics of interest include, but are not limited to, information processing and the theory and application of filtering, coding, transmitting, estimating, detecting, analyzing, recognizing, synthesizing, recording, and reproducing signals.