{"title":"EHC - k-NN:自适应k近邻的椭圆超复距离度量","authors":"Kaan Arik , Arzu Sürekçi , Hidayet Hüda Kösal","doi":"10.1016/j.neucom.2026.133051","DOIUrl":null,"url":null,"abstract":"<div><div>This study introduces a dimension-adaptive k-Nearest Neighbor (k-NN) model that employs a family of elliptic hypercomplex distance metrics, addressing the limitations of Euclidean geometry in heterogeneous and correlated data with tabular and image datasets. The approach reshapes the feature space using a negative real parameter <span><math><mi>p</mi><mo><</mo><mn>0</mn></math></span>, enabling curvature-controlled neighborhoods that better capture local structure. In the proposed method, each data instance is represented as an <span><math><mi>n</mi></math></span>-dimensional elliptic hypercomplex number, and distances are computed through a norm that re-weights even- and odd- indexed components depending on <span><math><mi>p</mi></math></span>. The proposed method is dimension-adaptive in the sense that each real-valued feature vector of length <span><math><mi>d</mi></math></span> is mapped to the smallest elliptic hypercomplex algebra of dimension <span><math><msup><mn>2</mn><mrow><mi>m</mi></mrow></msup></math></span> satisfying <span><math><msup><mn>2</mn><mrow><mi>m</mi><mo>−</mo><mn>1</mn></mrow></msup><mo><</mo><mi>d</mi><mo>≤</mo><msup><mn>2</mn><mrow><mi>m</mi></mrow></msup></math></span>. When <span><math><mi>d</mi><mo>≠</mo><msup><mn>2</mn><mrow><mi>m</mi></mrow></msup></math></span>, the remaining components are zero-padded, so distance computations are carried out consistently in the corresponding <span><math><msup><mn>2</mn><mrow><mi>m</mi></mrow></msup></math></span>-dimensional elliptic hypercomplex space. Experiments were conducted on five tabular UCI + two image-derived benchmarks selected for their diversity in feature types and class structure. Performance was evaluated using classification performance evaluation metrics under identical <span><math><mi>k</mi></math></span> settings. The proposed metric yields clear gains over Euclidean k-NN, particularly in <em>Wine</em> (approximately <span><math><mn>2.0</mn></math></span>-<span><math><mn>2.3</mn><mi>%</mi></math></span>) and <em>Breast Cancer</em> (approximately <span><math><mn>1.4</mn><mi>%</mi></math></span>). Improvements are moderate in <em>Car Evaluation</em>, while <em>Iris</em> and <em>Banknote Authentication</em> exhibit minimal change due to saturated separability and dominant attributes. On image-derived benchmarks (Seeds/Wheat and Image Segmentation), the proposed metric also delivers consistent improvements, typically around +2.0-2.7% in accuracy and +2.4-2.9% in F1-score compared with Euclidean k-NN. Further comparisons against metric-learning and manifold-inspired baselines (LMNN and geodesic distance) indicate that the proposed hypercomplex metric remains competitive and stable across neighborhood sizes, reinforcing its robustness beyond Euclidean geometry. Overall, the results indicate that the performance gains stem from the <span><math><mi>p</mi></math></span>-induced anisotropy of the elliptic hypercomplex norm, which reshapes neighborhood geometry to better align with heterogeneous and correlated feature structures, thus positioning elliptic hypercomplex k-NN as a robust alternative to Euclidean k-NN.</div></div>","PeriodicalId":19268,"journal":{"name":"Neurocomputing","volume":"676 ","pages":"Article 133051"},"PeriodicalIF":6.5000,"publicationDate":"2026-05-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"EHC k-NN: Elliptic hypercomplex distance metrics for dimension-adaptive k-nearest neighbor\",\"authors\":\"Kaan Arik , Arzu Sürekçi , Hidayet Hüda Kösal\",\"doi\":\"10.1016/j.neucom.2026.133051\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>This study introduces a dimension-adaptive k-Nearest Neighbor (k-NN) model that employs a family of elliptic hypercomplex distance metrics, addressing the limitations of Euclidean geometry in heterogeneous and correlated data with tabular and image datasets. The approach reshapes the feature space using a negative real parameter <span><math><mi>p</mi><mo><</mo><mn>0</mn></math></span>, enabling curvature-controlled neighborhoods that better capture local structure. In the proposed method, each data instance is represented as an <span><math><mi>n</mi></math></span>-dimensional elliptic hypercomplex number, and distances are computed through a norm that re-weights even- and odd- indexed components depending on <span><math><mi>p</mi></math></span>. The proposed method is dimension-adaptive in the sense that each real-valued feature vector of length <span><math><mi>d</mi></math></span> is mapped to the smallest elliptic hypercomplex algebra of dimension <span><math><msup><mn>2</mn><mrow><mi>m</mi></mrow></msup></math></span> satisfying <span><math><msup><mn>2</mn><mrow><mi>m</mi><mo>−</mo><mn>1</mn></mrow></msup><mo><</mo><mi>d</mi><mo>≤</mo><msup><mn>2</mn><mrow><mi>m</mi></mrow></msup></math></span>. When <span><math><mi>d</mi><mo>≠</mo><msup><mn>2</mn><mrow><mi>m</mi></mrow></msup></math></span>, the remaining components are zero-padded, so distance computations are carried out consistently in the corresponding <span><math><msup><mn>2</mn><mrow><mi>m</mi></mrow></msup></math></span>-dimensional elliptic hypercomplex space. Experiments were conducted on five tabular UCI + two image-derived benchmarks selected for their diversity in feature types and class structure. Performance was evaluated using classification performance evaluation metrics under identical <span><math><mi>k</mi></math></span> settings. The proposed metric yields clear gains over Euclidean k-NN, particularly in <em>Wine</em> (approximately <span><math><mn>2.0</mn></math></span>-<span><math><mn>2.3</mn><mi>%</mi></math></span>) and <em>Breast Cancer</em> (approximately <span><math><mn>1.4</mn><mi>%</mi></math></span>). Improvements are moderate in <em>Car Evaluation</em>, while <em>Iris</em> and <em>Banknote Authentication</em> exhibit minimal change due to saturated separability and dominant attributes. On image-derived benchmarks (Seeds/Wheat and Image Segmentation), the proposed metric also delivers consistent improvements, typically around +2.0-2.7% in accuracy and +2.4-2.9% in F1-score compared with Euclidean k-NN. Further comparisons against metric-learning and manifold-inspired baselines (LMNN and geodesic distance) indicate that the proposed hypercomplex metric remains competitive and stable across neighborhood sizes, reinforcing its robustness beyond Euclidean geometry. Overall, the results indicate that the performance gains stem from the <span><math><mi>p</mi></math></span>-induced anisotropy of the elliptic hypercomplex norm, which reshapes neighborhood geometry to better align with heterogeneous and correlated feature structures, thus positioning elliptic hypercomplex k-NN as a robust alternative to Euclidean k-NN.</div></div>\",\"PeriodicalId\":19268,\"journal\":{\"name\":\"Neurocomputing\",\"volume\":\"676 \",\"pages\":\"Article 133051\"},\"PeriodicalIF\":6.5000,\"publicationDate\":\"2026-05-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Neurocomputing\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S0925231226004480\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2026/2/11 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Neurocomputing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0925231226004480","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2026/2/11 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
EHC k-NN: Elliptic hypercomplex distance metrics for dimension-adaptive k-nearest neighbor
This study introduces a dimension-adaptive k-Nearest Neighbor (k-NN) model that employs a family of elliptic hypercomplex distance metrics, addressing the limitations of Euclidean geometry in heterogeneous and correlated data with tabular and image datasets. The approach reshapes the feature space using a negative real parameter , enabling curvature-controlled neighborhoods that better capture local structure. In the proposed method, each data instance is represented as an -dimensional elliptic hypercomplex number, and distances are computed through a norm that re-weights even- and odd- indexed components depending on . The proposed method is dimension-adaptive in the sense that each real-valued feature vector of length is mapped to the smallest elliptic hypercomplex algebra of dimension satisfying . When , the remaining components are zero-padded, so distance computations are carried out consistently in the corresponding -dimensional elliptic hypercomplex space. Experiments were conducted on five tabular UCI + two image-derived benchmarks selected for their diversity in feature types and class structure. Performance was evaluated using classification performance evaluation metrics under identical settings. The proposed metric yields clear gains over Euclidean k-NN, particularly in Wine (approximately -) and Breast Cancer (approximately ). Improvements are moderate in Car Evaluation, while Iris and Banknote Authentication exhibit minimal change due to saturated separability and dominant attributes. On image-derived benchmarks (Seeds/Wheat and Image Segmentation), the proposed metric also delivers consistent improvements, typically around +2.0-2.7% in accuracy and +2.4-2.9% in F1-score compared with Euclidean k-NN. Further comparisons against metric-learning and manifold-inspired baselines (LMNN and geodesic distance) indicate that the proposed hypercomplex metric remains competitive and stable across neighborhood sizes, reinforcing its robustness beyond Euclidean geometry. Overall, the results indicate that the performance gains stem from the -induced anisotropy of the elliptic hypercomplex norm, which reshapes neighborhood geometry to better align with heterogeneous and correlated feature structures, thus positioning elliptic hypercomplex k-NN as a robust alternative to Euclidean k-NN.
期刊介绍:
Neurocomputing publishes articles describing recent fundamental contributions in the field of neurocomputing. Neurocomputing theory, practice and applications are the essential topics being covered.