{"title":"Improved Perceptual Loss for Sketch Image Domain.","authors":"Chang Wook Seo","doi":"10.3390/jimaging11090323","DOIUrl":null,"url":null,"abstract":"<p><p>Traditional perceptual loss functions, primarily designed for photographic images, often perform poorly in the sketch domain due to significant differences in visual representation. To address this domain gap, we propose an improved perceptual loss specifically designed for sketch images. Our method fine-tunes a pre-trained VGG-16 model on the ImageNet-Sketch dataset while strategically replacing max-pooling layers with spatial and channel attention mechanisms. We comprehensively evaluate our approach across three dimensions: generation quality, sketch retrieval performance, and feature space organization. Experimental results demonstrate consistent improvements across all evaluation metrics, with our method achieving the best generation performance, over 10% improvement in sketch retrieval accuracy, and 6-fold improvement in class separability compared to baseline methods. The ablation studies confirm that both fine-tuning and attention mechanisms are essential components that work together effectively. Our domain-specific perceptual loss effectively bridges the gap between photographic and sketch domains, providing enhanced performance for various sketch-related computer vision applications, including generation, retrieval, and recognition.</p>","PeriodicalId":37035,"journal":{"name":"Journal of Imaging","volume":"11 9","pages":""},"PeriodicalIF":2.7000,"publicationDate":"2025-09-21","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC12470351/pdf/","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Journal of Imaging","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.3390/jimaging11090323","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q3","JCRName":"IMAGING SCIENCE & PHOTOGRAPHIC TECHNOLOGY","Score":null,"Total":0}
引用次数: 0
Abstract
Traditional perceptual loss functions, primarily designed for photographic images, often perform poorly in the sketch domain due to significant differences in visual representation. To address this domain gap, we propose an improved perceptual loss specifically designed for sketch images. Our method fine-tunes a pre-trained VGG-16 model on the ImageNet-Sketch dataset while strategically replacing max-pooling layers with spatial and channel attention mechanisms. We comprehensively evaluate our approach across three dimensions: generation quality, sketch retrieval performance, and feature space organization. Experimental results demonstrate consistent improvements across all evaluation metrics, with our method achieving the best generation performance, over 10% improvement in sketch retrieval accuracy, and 6-fold improvement in class separability compared to baseline methods. The ablation studies confirm that both fine-tuning and attention mechanisms are essential components that work together effectively. Our domain-specific perceptual loss effectively bridges the gap between photographic and sketch domains, providing enhanced performance for various sketch-related computer vision applications, including generation, retrieval, and recognition.