{"title":"Opto-mechanical design of Fourier transform spectral polarization imaging system based on dual Wollaston prisms","authors":"Jingping Zhu, Xiaofang Liu, Angze Li, Yue Liu, Xun Hou","doi":"10.1016/j.displa.2025.103026","DOIUrl":"10.1016/j.displa.2025.103026","url":null,"abstract":"<div><div>The integration of polarization and spectral imaging, which exploits the principle of spectral interference, has long been a subject of extensive research. However, existing literature lacks comprehensive discussions on the opto-mechanical design of Fourier transform spectral polarization imaging systems (FTSPIS) employing dual Wollaston prisms (DWP). We proposing a novel opto-mechanical design strategy for FTSPIS based on DWP. Specifically, the FTSPIS system is divided into three independent modules: telescopic objective lens with phase modulation module (PMM), eyepiece and splitting system with DWP, and secondary imaging system. For each module, a detailed optimization design process is presented, with special emphasis on the optical effects produced by each module and their compatibility with one another. A practical design instance is provided to illustrate the feasibility of constructing a visible-spectrum FTSPIS that can achieve a full field of view (FOV) of ± 5°. The effectiveness of the proposed design is verified through simulation analysis of interference patterns and demodulation. An elaboration on the mechanical structures for each module is also provided. These multi-degree-of-freedom adjustment capabilities enable precise control of the system’s performance, thus ensuring that the desired optical results can be achieved.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103026"},"PeriodicalIF":3.7,"publicationDate":"2025-03-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143684632","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-03-11DOI: 10.1016/j.displa.2025.103022
Ye-Rim Jeong , Eun Kyo Jung , Kook Chul Moon , Hwarim Im , Yong-Sang Kim
{"title":"Low-Power integrated driver circuits with multiple outputs based on Thin-Film transistors for mobile micro Light-Emitting diode displays","authors":"Ye-Rim Jeong , Eun Kyo Jung , Kook Chul Moon , Hwarim Im , Yong-Sang Kim","doi":"10.1016/j.displa.2025.103022","DOIUrl":"10.1016/j.displa.2025.103022","url":null,"abstract":"<div><div>This article proposes low-power integrated driver circuits based on low-temperature polycrystalline silicon (LTPS) thin-film transistors (TFTs) for small and medium-sized mobile micro light-emitting diode (μLED) displays. Two integrated driver circuits can output four signals to operate the μLED pixel circuit: the integrated scan driver outputs two scan signals, and the integrated emission driver outputs an emission signal and a sweep signal. Moreover, driver circuits are designed to enable the always-on-display (AOD) mode, which consumes less power than the normal mode. Compared to the normal mode, the proposed circuits do not output unnecessary signals in the AOD mode, reducing power consumption by 38.0 and 29.6% in the integrated scan and emission circuits. In addition, compared to the gate driver circuits with a single output signal. The proposed integrated scan and emission driver circuits reduce power consumption by 16.6 (25.2) and 42.5 (51.3)% in the normal (AOD) mode, respectively, by sharing the control unit for multiple outputs. We also fabricated the proposed driver circuits and verified that the proposed circuits successfully output multiple signals for mobile μLED pixel circuits.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103022"},"PeriodicalIF":3.7,"publicationDate":"2025-03-11","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143629179","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-03-10DOI: 10.1016/j.displa.2025.103024
N. Senyer , D. Durmus
{"title":"An AI-driven framework for perceived display spectra: The effects of dimming, observer age, and viewing distance","authors":"N. Senyer , D. Durmus","doi":"10.1016/j.displa.2025.103024","DOIUrl":"10.1016/j.displa.2025.103024","url":null,"abstract":"<div><div>Displays have become ubiquitous in modern society, serving as pervasive light sources that exert visual and non-visual effects on human physiology and behavior. Despite their widespread use and impact, a universal framework for characterizing perceived display light output across various viewing conditions still needs to be developed. This study introduces a novel, AI-driven framework for comprehensive perceived display light output characterization, accounting for the effects of observer age, viewing distance, and display dimming. The framework employs a deep neural network (DNN) trained on an extensive dataset of measured display spectra to predict spectral power distributions (SPDs) from RGB inputs. To simulate real-world scenarios, the DNN-predicted SPDs were transformed to account for viewing distance (36 cm–71 cm), display dimming (0–100 %), and observer age (1–100 years). The initial model achieved high accuracy (<em>R</em>2<sub>avg</sub> = 0.99), maintaining robust performance even for challenging cases (<em>R</em><sup>2</sup> > 0.94). Results show high accuracy in predicting photometric, colorimetric, and circadian measures. Future research will incorporate other parameters to the proposed framework.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103024"},"PeriodicalIF":3.7,"publicationDate":"2025-03-10","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143631880","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-03-06DOI: 10.1016/j.displa.2025.103016
Mengyao Qi , Kenta Ono , Lujin Mao , Makoto Watanabe , Jinghua Huang
{"title":"Hedonic or Utilitarian: The effect of video and text styles on consumers’ visual attention, mental imagery quality, and purchase intention in online food ordering interfaces","authors":"Mengyao Qi , Kenta Ono , Lujin Mao , Makoto Watanabe , Jinghua Huang","doi":"10.1016/j.displa.2025.103016","DOIUrl":"10.1016/j.displa.2025.103016","url":null,"abstract":"<div><div>Given the substantial resources allocated to video and text presentations, indiscriminately applying video and text presentations regardless of information style may not improve sales or enhance the consumers’ ordering experience, and can even lead to unnecessary resource waste. Therefore, this study aims to examine the effects of video and text styles (hedonic vs. utilitarian) on consumers’ visual attention and subjective assessments. The results indicate that the utilitarian text style could be a favorable choice for online food ordering interfaces, as it not only captures greater visual attention but also enhances users’ mental imagery quality and purchase intention. Additionally, the selection of video styles should be guided by the specific objectives of the context: the utilitarian video style appears to be a more effective option for encouraging purchasing behavior, as it is associated with significantly higher fixation counts, while the hedonic video style may be more conducive to enhancing user engagement and extending online duration, as indicated by its longer average fixation duration. These findings provide valuable insights into the effectiveness of short-form video and text across different information styles, contributing to the design of more user-friendly online food ordering interfaces.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103016"},"PeriodicalIF":3.7,"publicationDate":"2025-03-06","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143591643","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-03-04DOI: 10.1016/j.displa.2025.103019
Xingkai Wang , Bingyuan Wang , Lina Xu , Luwen Yu
{"title":"Tailored information display: Effects of background colour and line spacing on visual search across different character types – An eye-tracking study","authors":"Xingkai Wang , Bingyuan Wang , Lina Xu , Luwen Yu","doi":"10.1016/j.displa.2025.103019","DOIUrl":"10.1016/j.displa.2025.103019","url":null,"abstract":"<div><div>In the era of expanding digital information, optimizing display factors has become increasingly critical for human–computer interaction. However, the broader impacts of background colour and line spacing on interfaces with different character types, under controlled luminance contrast remain underexplored. This study investigates the effects of background colour and line spacing on search performance, eye-tracking metrics, and subjective reports across English, Chinese, and numerals in controlled luminance conditions using a pseudo-text visual search task. Findings reveal the distinct impacts and mechanisms for each character type. Red background significantly hindered search performance, particularly in English, while cyan improved accuracy and subjective ratings across all characters. Colour difference emerged as a critical determinant of visual comfort, and its influence is moderated by pupil diameter. Double line spacing provided an optimal balance between search performance and minimizing eye movement effort, further increases in line spacing impair search speed, especially in numerals. Fixation duration and fixation count mediated the effects of line spacing on legibility in numerals and English, respectively. These findings offer actionable insights for optimizing interfaces with different character types and highlight the need to explore further how display factors contribute to visual experience in complex scripts.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103019"},"PeriodicalIF":3.7,"publicationDate":"2025-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143591642","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-03-04DOI: 10.1016/j.displa.2025.103020
Yijia Wen , Xiaoyan Luo , Lei Zheng , Liangnan Qi , Xiaofeng Shi
{"title":"vAQA-SS: Vision-based action quality assessment for style-based skiing","authors":"Yijia Wen , Xiaoyan Luo , Lei Zheng , Liangnan Qi , Xiaofeng Shi","doi":"10.1016/j.displa.2025.103020","DOIUrl":"10.1016/j.displa.2025.103020","url":null,"abstract":"<div><div>Vision-based Action Quality Assessment (AQA) aims to evaluate action quality in video data, aligning with the subjective scores of human experts. Due to the unique challenges posed by different sports, it is difficult to design a uniform AQA system applicable to all sports. Consequently, many current sports AQA methods focus on specific disciplines such as diving and gymnastics. In contrast, skiing AQA, characterized by high-dynamic actions and complex outdoor scenes, faces additional challenges. Therefore, we constructed a specific dataset for style-based skiing, which focuses athlete’s movement style and execution, encompassing diverse skiing events with detailed annotations on action classes and athletes’ final scores, named Skiing-6. Leveraging this dataset, we designed two vision-based skiing action quality assessment (vAQA-SS) models. One model directly generates an absolute AQA score by measuring the quality of an athlete’s actions in the input video without any external reference, termed <em>or</em>AQA, which assesses athlete performance based on low-level spatiotemporal features of the video data alongside high-level pose features. The other model calculates a relative AQA score, deriving the performance score of an athlete’s actions from the source input video with a reference video, termed <em>wr</em>AQA. Finally, we conducted extensive experiments on Skiing-6 and SkiTB to demonstrate the effectiveness of our vAQA-SS models. The results demonstrate that our approach achieves significant improvements in both absolute evaluation (<em>or</em>AQA) and relative evaluation (<em>wr</em>AQA), surpassing other similar sports AQA methods.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103020"},"PeriodicalIF":3.7,"publicationDate":"2025-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143642249","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-02-27DOI: 10.1016/j.displa.2025.102992
Shuchang Zhou , Hanxin Wang , Qingbo Wu , Fanman Meng , Linfeng Xu , Wei Zhang , Hongliang Li
{"title":"Adversarially Regularized Tri-Transformer Fusion for continual multimodal egocentric activity recognition","authors":"Shuchang Zhou , Hanxin Wang , Qingbo Wu , Fanman Meng , Linfeng Xu , Wei Zhang , Hongliang Li","doi":"10.1016/j.displa.2025.102992","DOIUrl":"10.1016/j.displa.2025.102992","url":null,"abstract":"<div><div>Continual egocentric activity recognition aims to understand first-person activity from the multimodal data captured from wearable devices in streaming environments. Existing continual learning (CL) methods hardly acquire discriminative multimodal representations of activity classes from different isolated stages. To address this issue, this paper proposes an Adversarially Regularized Tri-Transformer Fusion (ARTF) model composed of three frozen transformer backbones with dynamic expansion architecture, which enables flexible and progressive multimodal representation fusion in the CL setting. To mitigate the confusion across different stages, we adopt an adversary-based confusion feature generation strategy to augment unknown classes, explicitly simulating out-stage features that closely resemble those within the stage. Then, the discriminative multimodal fusion representations could be learned by joint training on the current and augmented data at different stages. Experiments show that our model significantly outperforms state-of-the-art CL methods for multimodal continual egocentric activity recognition.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 102992"},"PeriodicalIF":3.7,"publicationDate":"2025-02-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143529734","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-02-26DOI: 10.1016/j.displa.2025.103007
Jiale Chao, Jialin Lei, Xionghui Zhou, Le Xie
{"title":"A general and flexible point cloud simplification method based on feature fusion","authors":"Jiale Chao, Jialin Lei, Xionghui Zhou, Le Xie","doi":"10.1016/j.displa.2025.103007","DOIUrl":"10.1016/j.displa.2025.103007","url":null,"abstract":"<div><div>Large-scale, high-density point cloud data often pose challenges for direct application in various downstream tasks. To address this issue, this paper introduces a flexible point cloud simplification method based on feature fusion. After conducting a comprehensive analysis of the input point cloud, the method fuses the density feature that reflects point cloud uniformity with local geometric features that capture shape details. Based on the simplification objectives and fused feature values, the method optimizes the point distribution from a global perspective. Subsequently, by removing distance factors, purely local geometric features are incorporated into the farthest point sampling process and a feature-weighted voxel farthest point sampling algorithm is proposed to prioritize the preservation of local feature points. With a refined mechanism for adjusting point numbers, the method finally achieves fast and reasonable simplification of massive point clouds. Furthermore, extensive experiments have been designed to explore the impact of the features involved and their sensitivity to simplification results, offering detailed recommendations for parameter configuration. This method supports flexible transitions between global uniformity and heavy local feature preservation. Comparative results with previous studies demonstrate its excellent balance, exhibiting strong competitiveness in both output point cloud quality and computational efficiency. The core source code is publicly available at: <span><span>https://github.com/chaojiale/PointCloudSimplification</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103007"},"PeriodicalIF":3.7,"publicationDate":"2025-02-26","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143529733","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-02-25DOI: 10.1016/j.displa.2025.103003
Lianmin Zhang , Hongkui Wang , Qionghua Luo , Wei Zhang , Haibing Yin , Tiansong Li , Li Yu , Wenyao Zhu
{"title":"Bayesian generation based foveated JND estimation in the DCT domain","authors":"Lianmin Zhang , Hongkui Wang , Qionghua Luo , Wei Zhang , Haibing Yin , Tiansong Li , Li Yu , Wenyao Zhu","doi":"10.1016/j.displa.2025.103003","DOIUrl":"10.1016/j.displa.2025.103003","url":null,"abstract":"<div><div>The Just Noticeable Distortion (JND) threshold refers to the inability of the human visual system (HVS) to perceive pixel changes below a certain visibility threshold. In this paper, we focus on the cross-domain operation problem of JND estimation in the DCT domain. In order to solve this problem and improve the accuracy of DCT-JND estimation, we design an autoregressive model based on the Bayesian generation theory to simulate the spontaneous predictive behavior of HVS. Based on this model, an entropy masking (EM) effect based JND moderator is then proposed. Considering the visual attention and foveated masking (VFM) effect, this paper predicts visual saliency and the fixation points in the DCT domain, an enhanced foveated masking effect based JND moderator is then presented. Finally, combined with other JND moderators, the Bayesian generation based foveated DCT-JND model is obtained. Subjective and objective experimental results show that the proposed model could further improve the accuracy of JND threshold estimation in the DCT domain while avoiding the cross-domain operation.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103003"},"PeriodicalIF":3.7,"publicationDate":"2025-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143510569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-02-25DOI: 10.1016/j.displa.2025.103002
Hui Hu , Yunhui Shi , Jin Wang , Nam Ling , Baocai Yin
{"title":"Feature enhanced spherical transformer for spherical image compression","authors":"Hui Hu , Yunhui Shi , Jin Wang , Nam Ling , Baocai Yin","doi":"10.1016/j.displa.2025.103002","DOIUrl":"10.1016/j.displa.2025.103002","url":null,"abstract":"<div><div>It is well known that the wide field of view of spherical images requires high resolution, which increases the challenges of storage and transmission. Recently, a spherical learning-based image compression method called OSLO has been proposed, which leverages HEALPix’s approximately uniform spherical sampling. However, HEALPix sampling can only utilize a fixed 3 × 3 convolution kernel, resulting in a limited receptive field and an inability to capture non-local information. This limitation hinders redundancy removal during the transform and texture synthesis during the inverse transform. To address this issue, we propose a feature-enhanced spherical Transformer-based image compression method that leverages HEALPix’s hierarchical structure. Specifically, to reduce the computational complexity of the Transformer’s attention mechanism, we divide the sphere into multiple windows using HEALPix’s hierarchical structure and compute attention within these spherical windows. Since there is no communication between adjacent windows, we introduce spherical convolution to aggregate information from neighboring windows based on their local correlation. Additionally, to enhance the representational ability of features, we incorporate an inverted residual bottleneck module for feature embedding and a feedforward neural network. Experimental results demonstrate that our method outperforms OSLO, achieving lower codec time.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"88 ","pages":"Article 103002"},"PeriodicalIF":3.7,"publicationDate":"2025-02-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"143510571","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}