{"title":"Deep analysis on the color language in film and television animation works via semantic segmentation technique","authors":"Lu Huang , Wendan Yang","doi":"10.1016/j.entcom.2025.100990","DOIUrl":null,"url":null,"abstract":"<div><div>In film and television animation works, colour is an ideographic symbol distinct from other elements. Color language and image language are separate concepts, with image language unable to fully capture the ideographic nature of color. Even when images appear visually similar, differences in brightness can create significant contrasts or opposing meanings. Thus, color language plays an ideographic role between color and concept. This paper uses semantic segmentation techniques to integrate the three modalities of color content and text, establishing a consistent representation relationship. We propose a color-depth (RGB-D) image semantic segmentation method based on two-stream weighted Gabor convolutional network fusion. To capture Orientation- and scale-invariant features, we design a weighted Gabor orientation filter within a deep convolutional network (DCN) to adapt to changes in Orientation and scale. A wide residual-weighted Gabor convolutional network extracts features from the dual-stream images of colour and depth. To quantitatively assess our method’s representational ability regarding the ideographic functions of color language, we conduct extensive experiments on public datasets. The results demonstrate that the proposed algorithm outperforms existing RGB-D image semantic segmentation methods. Specifically, the technique achieves superior accuracy across several performance metrics, with a notable improvement of 2.5%–6.6% compared to baseline models. Our approach enhances segmentation precision for objects of varying scales and directions. It exhibits robustness in complex lighting environments, thus confirming its potential in real-world applications of color language in animation.</div></div>","PeriodicalId":55997,"journal":{"name":"Entertainment Computing","volume":"55 ","pages":"Article 100990"},"PeriodicalIF":2.4000,"publicationDate":"2025-06-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Entertainment Computing","FirstCategoryId":"94","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1875952125000709","RegionNum":3,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"COMPUTER SCIENCE, CYBERNETICS","Score":null,"Total":0}
引用次数: 0
Abstract
In film and television animation works, colour is an ideographic symbol distinct from other elements. Color language and image language are separate concepts, with image language unable to fully capture the ideographic nature of color. Even when images appear visually similar, differences in brightness can create significant contrasts or opposing meanings. Thus, color language plays an ideographic role between color and concept. This paper uses semantic segmentation techniques to integrate the three modalities of color content and text, establishing a consistent representation relationship. We propose a color-depth (RGB-D) image semantic segmentation method based on two-stream weighted Gabor convolutional network fusion. To capture Orientation- and scale-invariant features, we design a weighted Gabor orientation filter within a deep convolutional network (DCN) to adapt to changes in Orientation and scale. A wide residual-weighted Gabor convolutional network extracts features from the dual-stream images of colour and depth. To quantitatively assess our method’s representational ability regarding the ideographic functions of color language, we conduct extensive experiments on public datasets. The results demonstrate that the proposed algorithm outperforms existing RGB-D image semantic segmentation methods. Specifically, the technique achieves superior accuracy across several performance metrics, with a notable improvement of 2.5%–6.6% compared to baseline models. Our approach enhances segmentation precision for objects of varying scales and directions. It exhibits robustness in complex lighting environments, thus confirming its potential in real-world applications of color language in animation.
期刊介绍:
Entertainment Computing publishes original, peer-reviewed research articles and serves as a forum for stimulating and disseminating innovative research ideas, emerging technologies, empirical investigations, state-of-the-art methods and tools in all aspects of digital entertainment, new media, entertainment computing, gaming, robotics, toys and applications among researchers, engineers, social scientists, artists and practitioners. Theoretical, technical, empirical, survey articles and case studies are all appropriate to the journal.