DisplaysPub Date : 2025-09-18DOI: 10.1016/j.displa.2025.103220
Xinyue Chen , Hang Shi , Shi-bing Guan , Wei Shao
{"title":"IP-MIML: A multi-instance multi-label learning framework for predicting protein subcellular localization from biological images","authors":"Xinyue Chen , Hang Shi , Shi-bing Guan , Wei Shao","doi":"10.1016/j.displa.2025.103220","DOIUrl":"10.1016/j.displa.2025.103220","url":null,"abstract":"<div><div>Recent studies indicate that the localization of proteins within a cell is essential for determining their functions and gaining insights into various cellular processes. With advances in microscopic imaging, accurate classification of bioimage-based protein subcellular localization patterns has attracted as much attention as ever. However, most bioimage-based protein subcellular location predictors are designed to allocate the protein image to one location, which overlooks the case that a protein may colocalize in different cellular compartments that deserve special attention. On the other hand, we could observe a protein expressed in multiple biological images derived from different tissues, it is still a challenge to summarize the localization patterns of that protein across all related images. Based on the above considerations, we propose a multi-instance multi-label learning framework to determine the subcellular localization of proteins from biological images (<em>i.e.,</em> IP-MIML). Specifically, we first treat one protein as a bag and all images belonging to it as instances and introduce the self-attention mechanism to learn instance-level representation by considering their correlations. Then, a bag-concept layer is developed to discover the latent relation between the inputs and the output semantic labels. In addition, we also incorporate an optimal transport (OT) based formulation to learn the label distribution and exploit label correlations, simultaneously. Finally, a dynamic threshold method is utilized for adjusting the multi-label prediction results. We evaluated our method on normal and cancer protein bioimages, and the experimental results indicate that the proposed IP-MIML not only can achieve higher accuracy in predicting the cellular compartments of proteins with multiple localizations, but also can detect potential cancer biomarker proteins that have significant localization differences between normal and cancer tissues.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"91 ","pages":"Article 103220"},"PeriodicalIF":3.4,"publicationDate":"2025-09-18","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145219942","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-09-17DOI: 10.1016/j.displa.2025.103227
Yang Zhu , Yong-Cheng Lin
{"title":"LiMS-Net: Lightweight metal surface defect detection network","authors":"Yang Zhu , Yong-Cheng Lin","doi":"10.1016/j.displa.2025.103227","DOIUrl":"10.1016/j.displa.2025.103227","url":null,"abstract":"<div><div>This study aims to enhance the accuracy of detecting defects on metal surfaces by proposing a lightweight metal surface defect detection network (LiMS-Net). The backbone of LiMS-Net incorporates a residual synchronous convolutional block feature extraction module that utilizes multi-scale convolution kernels. Features are concurrently processed using these multi-scale convolution kernels. In the neck stage, a Conv-MLP module that extracts global image features. This module is further enhanced by shift operations that improve information interaction among different regions of the features. To further enhance feature interaction across different scales and improve detection accuracy, a cross-scale feature fusion block is proposed. This approach alleviates feature loss issues caused by extensive feature processing. This study employed the advanced object detection methods and conducted comparative experiments using publicly available defect databases. Compared to the advanced object detection methods, LiMS-Net demonstrated superior performance across all databases while utilizing fewer parameters.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"91 ","pages":"Article 103227"},"PeriodicalIF":3.4,"publicationDate":"2025-09-17","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145108689","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-09-16DOI: 10.1016/j.displa.2025.103228
Julie Fournier , Pauline Mann , Emma Bucher , Elise Etchamendy , Giorgia Tiscini , Meriem Outtas , Lu Zhang , Myriam Chérel
{"title":"Investigating the influence of affinity on the gaze behavior of individuals with Autism Spectrum Disorder (ASD) based on their unique autistic traits","authors":"Julie Fournier , Pauline Mann , Emma Bucher , Elise Etchamendy , Giorgia Tiscini , Meriem Outtas , Lu Zhang , Myriam Chérel","doi":"10.1016/j.displa.2025.103228","DOIUrl":"10.1016/j.displa.2025.103228","url":null,"abstract":"<div><div>Autism spectrum is very wide, but autistic people share some traits such as social or language impairments. Another common characteristic is a strong passion, called an affinity, which can be anything from a movie character to a topic like history, or even a specific object. Numerous testimonies attest to the support provided by affinities for autistic individuals, offering a reassuring space in a sometimes frightening world. Clinical psychologists consider affinities as keys that can unlock language and learning for people with ASD. However, objective evidence of this role is still lacking. In this study, we used eye-tracking technology to explore the visual attention patterns of autistic individuals when presented with their affinity compared to neutral stimuli. We recruited 52 autistic participants and showed them 38 images: 10 featuring their affinity and 28 neutral ones, while recording their eye movements. Eye-tracking data provided crucial insights into how visual attention is modulated by affinities. Our results reveal significant variability in visual engagement depending on the specific autistic traits and affinities of the participants. Some showed heightened visual engagement with affinity images, while others withdrew their gaze. Some exhibited a mixed response, with both increased engagement and gaze withdrawal, and a few showed no difference between the two sets of images. These findings highlight the complex relationship between visual attention and affinities in autistic individuals, highlighting the potential of eye-tracking as a tool for understanding and leveraging these affinities in therapeutic and educational settings.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"91 ","pages":"Article 103228"},"PeriodicalIF":3.4,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145219944","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-09-16DOI: 10.1016/j.displa.2025.103200
Zaidao Han , Sanqian Li , Xiang Wang , Xiaoqing Hu , Risa Higashita , Jiang Liu
{"title":"Multi-path sensory substitution device navigates the blind and visually impaired individuals","authors":"Zaidao Han , Sanqian Li , Xiang Wang , Xiaoqing Hu , Risa Higashita , Jiang Liu","doi":"10.1016/j.displa.2025.103200","DOIUrl":"10.1016/j.displa.2025.103200","url":null,"abstract":"<div><div>With the rapid advancements in computer vision technology, real-time autonomous navigation systems for blind and visually impaired individuals (BVIs) leveraging scene understanding have become increasingly feasible. Remarkably, existing navigation systems demonstrate excellent performance in obstacle detection and directional guidance for BVIs. Nevertheless, they lack the spatial perception of the entire scene for BVIs to reach the autonomous navigation. To overcome this issue, we propose a novel computer vision-based method to generate Multiple Paths with Sensory Substitution Device (MP-SSD) system, aiming to effectively and conveniently provide key autonomous navigation information for BVIs in outdoor routes. The MP-SSD system combines potential navigation target detection, path planning, and 3D Semantic Scene Completion (SSC) techniques to develop a widely applicable environmental detection and sensory substitution device (SSD), that can cater to the practical requirements of BVIs. Specifically, MP-SSD can extract semantic and spatial information from the input RGB image through the three-dimensional SSC model, and complete the information of the invisible area of the scene, thereby identifying potential target points and planning the shortest navigation path. During the interaction process, the system provides multi-path prompts through spatial audio to ensure accurate guidance. Experimental analyses on experience feedback of BVIs indicate that MP-SSD can effectively help BVIs actively acquire valuable navigation information in outdoor environments, thereby enhancing their autonomous mobility.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"91 ","pages":"Article 103200"},"PeriodicalIF":3.4,"publicationDate":"2025-09-16","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145095570","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-09-15DOI: 10.1016/j.displa.2025.103221
Li Xu , Yingjie Zhou , Sitong Liu , Farong Wen , Yu Zhou , Xiaohong Liu , Jie Guo , Yu Wang , Jiezhang Cao
{"title":"Facial quality assessment of digital humans: A dual-branch framework integrating morphological harmony and expressive coordination","authors":"Li Xu , Yingjie Zhou , Sitong Liu , Farong Wen , Yu Zhou , Xiaohong Liu , Jie Guo , Yu Wang , Jiezhang Cao","doi":"10.1016/j.displa.2025.103221","DOIUrl":"10.1016/j.displa.2025.103221","url":null,"abstract":"<div><div>With the rapid advancement of metaverse technologies, digital humans (DH), as core interactive entities in virtual-physical integrated ecosystems, face unique challenges in their quality assessment frameworks. Existing research predominantly focuses on quantifying natural image distortions but fails to address DH-specific issues such as facial morphological disharmony and expression incoherence. To bridge this gap, we propose a dual-branch quality assessment framework for digital humans: (1) Leveraging medical aesthetic priors, we construct structural features based on facial aesthetic subunits and model temporal dependencies using gated recurrent units, combined with a loss-averse pooling strategy to capture transient severe distortions. (2) We quantify expression coordination through multi-dimensional Action Unit (AU) topology graphs, proposing triple-edge definitions and regressing dynamic distortion levels via graph convolutional networks. Experiments on the multiple THQA datasets demonstrate that our framework significantly outperforms conventional methods in subjective mean opinion score consistency, with the dynamic branch playing a dominant role in performance optimization. This work establishes a quantifiable evaluation standard for DH modeling refinement and real-time rendering.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"91 ","pages":"Article 103221"},"PeriodicalIF":3.4,"publicationDate":"2025-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145157454","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-09-15DOI: 10.1016/j.displa.2025.103219
Jiahui Zhu , Aoqun Jian , Le Yang , RunFang Hao , Luxiao Sang , Yang Ge , Rihui Kang , Shengbo Sang
{"title":"Spatial derivative-guided SNR regional differentiation enhancement fusion strategy for low-light image enhancement","authors":"Jiahui Zhu , Aoqun Jian , Le Yang , RunFang Hao , Luxiao Sang , Yang Ge , Rihui Kang , Shengbo Sang","doi":"10.1016/j.displa.2025.103219","DOIUrl":"10.1016/j.displa.2025.103219","url":null,"abstract":"<div><div>Low-light image enhancement aims to improve brightness and contrast while preserving image content. Research into this problem has made significant progress with the development of deep learning technology. However, the Signal-to-Noise Ratio(SNR) of different regions varies greatly when processing images with drastic changes in brightness. Existing methods often produce artifacts and noise that degrade image quality. To address these problems,the proposed method incorporates local and global prior knowledge into the image, employing an efficient local-to-local and local-to-global feature fusion mechanism. This facilitates the generation of enhanced images that exhibit enhanced naturalness and a broader color dynamic range. In this approach, a spatial derivative-guided SNR regional differentiation enhancement fusion strategy is introduced. The enhancement of low SNR regions is processed in the frequency domain using the Fast Fourier Transform (FFT) while the enhancement of high/normal SNR regions is handled by a convolutional encoder. The convolution residual block structure, which captures local information, generates short-range branches. The FFT module in the frequency domain generates long-range branches. The fusion of the two is guided by the SNR information of the original image. This approach also incorporates spatial derivatives as local priors in a low-light image enhancement network with an encoder–decoder structure. The encoder employs the symmetrical properties of the image’s spatial derivatives and incorporates correlating modules for the suppression of noise. Experiments conducted on disparate datasets illustrate that our approach outperforms existing state-of-the-art(SOTA) methods in terms of visual quality. Furthermore, the single-frame inference time can be reduced to 0.079 s.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"91 ","pages":"Article 103219"},"PeriodicalIF":3.4,"publicationDate":"2025-09-15","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145095568","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-09-13DOI: 10.1016/j.displa.2025.103217
Zhiyong Deng, Ronggui Wang, Lixia Xue, Juan Yang
{"title":"AsPrompt: Attribute-structured knowledge-guided dual-modal coupling prompt learning for few-shot image classification","authors":"Zhiyong Deng, Ronggui Wang, Lixia Xue, Juan Yang","doi":"10.1016/j.displa.2025.103217","DOIUrl":"10.1016/j.displa.2025.103217","url":null,"abstract":"<div><div>The few-shot image classification task involves classifying images when only a limited number of training images are available. This field has seen significant advancements in recent years due to the development of pre-trained vision-language models (e.g., CLIP), which exhibit strong generalization capabilities. Recent studies have further leveraged classes-related descriptions as part of prompt learning to better adapt these foundational vision-language models for downstream tasks. However, the textual descriptions used in traditional methods often lack sufficient class-discriminative information, limiting the model’s expressiveness on unseen data domains. Given that large language models possess rich structured knowledge bases, they offer new avenues for enhancing textual information. Against this backdrop, we propose a novel method called AsPrompt, which integrates attribute-structured knowledge guidance with a dual-modal coupling prompt learning mechanism. This approach not only enriches class-discriminative textual information but also effectively integrates structured knowledge with traditional textual information by capturing the structured relationships between entity sets and attribute sets. Experimental results demonstrate that AsPrompt surpasses other state-of-the-art prompt learning methods on 11 different few-shot image classification datasets, showcasing its superior performance. The code can be found at <span><span>https://github.com/SandyPrompt/AsPrompt</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"91 ","pages":"Article 103217"},"PeriodicalIF":3.4,"publicationDate":"2025-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145095573","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-09-13DOI: 10.1016/j.displa.2025.103218
Zefeng Ying , Shuqi wang , Ping Shi , Xiumei Jia
{"title":"PDCR-SR: Enhancing facial super-resolution with multi-scale prior dictionary and region-specific contrastive regularization","authors":"Zefeng Ying , Shuqi wang , Ping Shi , Xiumei Jia","doi":"10.1016/j.displa.2025.103218","DOIUrl":"10.1016/j.displa.2025.103218","url":null,"abstract":"<div><div>Facial super-resolution involves reconstructing high-quality facial images from low-resolution face images and restoring rich facial details. Existing algorithms often struggle with the restoration of global structural details and localized facial features. To address these challenges, we propose the PDCR-SR method, which introduces a Multi-Scale Prior Dictionary (MSPD) for leveraging high-quality features across scales, enhancing detail reconstruction. Additionally, the Region-Specific Contrastive Regularization Module (RSCR) focuses on improving the texture and accuracy of localized areas such as skin, eyes, nose, and mouth. Extensive comparison results prove that our model has better reconstruction performance on both synthetic faces and real wild faces, superior to other existing methods in terms of quantitative indicators and visual quality.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"91 ","pages":"Article 103218"},"PeriodicalIF":3.4,"publicationDate":"2025-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145095572","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-09-13DOI: 10.1016/j.displa.2025.103222
Francisco Felip-Miralles, Julia Galán Serrano, Almudena Palacios-Ibáñez
{"title":"Can different contexts shape the perception of a product? A study using eye-tracking in virtual environments","authors":"Francisco Felip-Miralles, Julia Galán Serrano, Almudena Palacios-Ibáñez","doi":"10.1016/j.displa.2025.103222","DOIUrl":"10.1016/j.displa.2025.103222","url":null,"abstract":"<div><div>Virtual reality (VR) environments offer immersive experiences that improve products presentation and evaluation by allowing realistic representations and a more accurate interaction. Advances in both hardware and software have made this tool popular. It also improves visual quality for incorporating functions like eye tracking and enables user behaviour to be analysed in more depth. Despite all this favouring its application while evaluating products, it is still necessary to investigate how different factors influence perceptions of virtual prototypes (VP) to make the most of the advantages of VR. The present research uses three case studies to explore how a context (neutral, natural, urban) impacts evaluations of product characteristics, emotional response, trust in response, observation patterns, behaviour in virtual environments, cybersickness levels and feeling of presence. The results reveal that the context does not significantly influence product evaluation, but impacts emotional response (more positive on the natural vs. urban background) and form of observation (VP is observed longer when presented on a neutral background). These findings open up new opportunities to optimise products design and evaluation. However, future research should consider other variables like users’ age and other product categories.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"91 ","pages":"Article 103222"},"PeriodicalIF":3.4,"publicationDate":"2025-09-13","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145060460","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}
DisplaysPub Date : 2025-09-12DOI: 10.1016/j.displa.2025.103215
Huanqing Yan , Bo Sun , Jun He
{"title":"Uncertainty-aware weakly supervised temporal action localization with knowledge selection","authors":"Huanqing Yan , Bo Sun , Jun He","doi":"10.1016/j.displa.2025.103215","DOIUrl":"10.1016/j.displa.2025.103215","url":null,"abstract":"<div><div>Weakly-Supervised Temporal Action Localization (WS-TAL) aims to localize actions in untrimmed videos using only video-level labels. The core challenge is the lack of fine-grained annotations, which leads to high prediction uncertainty and confusion between actions and background. To address this, we propose an <em>Uncertainty-Aware and Knowledge-Selection</em> (UAKS) approach. Specifically, we integrate two uncertainty estimation strategies to cooperatively optimize the model and leverage uncertainty to guide external knowledge selection. First, evidential learning estimates model uncertainty, generating more confident predictions via regularization. Second, probabilistic distribution learning captures data uncertainty. Both uncertainties jointly guide model optimization. Additionally, uncertainty-driven knowledge selection enables the efficient utilization of external knowledge under weak supervision. Experiments show that our method improves accuracy and robustness, with 12.9% and 2% accuracy improvements on THUMOS and ActivityNet v1.3 datasets respectively, demonstrating the potential of uncertainty modeling in WS-TAL.</div></div>","PeriodicalId":50570,"journal":{"name":"Displays","volume":"91 ","pages":"Article 103215"},"PeriodicalIF":3.4,"publicationDate":"2025-09-12","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":null,"resultStr":null,"platform":"Semanticscholar","paperid":"145095569","PeriodicalName":null,"FirstCategoryId":null,"ListUrlMain":null,"RegionNum":2,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":"","EPubDate":null,"PubModel":null,"JCR":null,"JCRName":null,"Score":null,"Total":0}