Dual-stream temporal–spectral framework for accurate eye movement event detection

IF 3.6 2区 工程技术 Q2 ENGINEERING, ELECTRICAL & ELECTRONIC
Yang Zheng , Zhiyong Yu , Ruimin Li , Jimin Liang , Kaitai Guo
{"title":"Dual-stream temporal–spectral framework for accurate eye movement event detection","authors":"Yang Zheng ,&nbsp;Zhiyong Yu ,&nbsp;Ruimin Li ,&nbsp;Jimin Liang ,&nbsp;Kaitai Guo","doi":"10.1016/j.sigpro.2025.110252","DOIUrl":null,"url":null,"abstract":"<div><div>Accurate detection of eye movement events is crucial for unraveling complex visual behaviors and driving advancements in neuroscience and cognitive research. Deep learning approaches have demonstrated exceptional potential, surpassing traditional and machine learning methods by effectively capturing and modeling intricate patterns in eye movement data. However, existing deep learning methods face limitations, with small receptive fields struggling to capture cross-boundary information and large receptive fields failing to precisely delineate event borders. To address this, we propose GazeFusion, a novel dual-stream framework that integrates global temporal and local spectral features for comprehensive eye movement event detection. The framework comprises a Mamba Temporal Feature Extractor for long-range temporal dependencies, a CNN Spectral Feature Extractor for explicit frequency-domain representation, a Contextualizer leveraging attention mechanisms for global–local temporal–spectral feature fusion, and a Sequential Enhancer for refined sequential modeling. Experimental results on three publicly available datasets consistently demonstrate that GazeFusion outperforms state-of-the-art methods across all categories, offering superior accuracy and robust boundary detection.</div></div>","PeriodicalId":49523,"journal":{"name":"Signal Processing","volume":"239 ","pages":"Article 110252"},"PeriodicalIF":3.6000,"publicationDate":"2025-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Signal Processing","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S0165168425003664","RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, ELECTRICAL & ELECTRONIC","Score":null,"Total":0}
引用次数: 0

Abstract

Accurate detection of eye movement events is crucial for unraveling complex visual behaviors and driving advancements in neuroscience and cognitive research. Deep learning approaches have demonstrated exceptional potential, surpassing traditional and machine learning methods by effectively capturing and modeling intricate patterns in eye movement data. However, existing deep learning methods face limitations, with small receptive fields struggling to capture cross-boundary information and large receptive fields failing to precisely delineate event borders. To address this, we propose GazeFusion, a novel dual-stream framework that integrates global temporal and local spectral features for comprehensive eye movement event detection. The framework comprises a Mamba Temporal Feature Extractor for long-range temporal dependencies, a CNN Spectral Feature Extractor for explicit frequency-domain representation, a Contextualizer leveraging attention mechanisms for global–local temporal–spectral feature fusion, and a Sequential Enhancer for refined sequential modeling. Experimental results on three publicly available datasets consistently demonstrate that GazeFusion outperforms state-of-the-art methods across all categories, offering superior accuracy and robust boundary detection.
精确眼动事件检测的双流时间-光谱框架
准确检测眼球运动事件对于揭示复杂的视觉行为和推动神经科学和认知研究的进步至关重要。深度学习方法通过有效地捕获和建模眼动数据中的复杂模式,超越了传统和机器学习方法,显示出了非凡的潜力。然而,现有的深度学习方法面临局限性,小的接受场难以捕获跨边界信息,大的接受场无法精确描绘事件边界。为了解决这个问题,我们提出了一种新的双流框架GazeFusion,它集成了全局时间和局部光谱特征,用于全面的眼动事件检测。该框架包括用于远程时间依赖性的曼巴时间特征提取器,用于显式频域表示的CNN频谱特征提取器,用于全局-局部时间-频谱特征融合的利用注意机制的上下文化器,以及用于精细顺序建模的顺序增强器。在三个公开可用的数据集上的实验结果一致表明,GazeFusion在所有类别中都优于最先进的方法,提供卓越的准确性和鲁棒性边界检测。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Signal Processing
Signal Processing 工程技术-工程:电子与电气
CiteScore
9.20
自引率
9.10%
发文量
309
审稿时长
41 days
期刊介绍: Signal Processing incorporates all aspects of the theory and practice of signal processing. It features original research work, tutorial and review articles, and accounts of practical developments. It is intended for a rapid dissemination of knowledge and experience to engineers and scientists working in the research, development or practical application of signal processing. Subject areas covered by the journal include: Signal Theory; Stochastic Processes; Detection and Estimation; Spectral Analysis; Filtering; Signal Processing Systems; Software Developments; Image Processing; Pattern Recognition; Optical Signal Processing; Digital Signal Processing; Multi-dimensional Signal Processing; Communication Signal Processing; Biomedical Signal Processing; Geophysical and Astrophysical Signal Processing; Earth Resources Signal Processing; Acoustic and Vibration Signal Processing; Data Processing; Remote Sensing; Signal Processing Technology; Radar Signal Processing; Sonar Signal Processing; Industrial Applications; New Applications.
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信