Hosna Darvishi, R. Azmi, Fatemeh Moradian, Maral Zarvani
{"title":"Fashion Compatibility Learning Via Triplet-Swin Transformer","authors":"Hosna Darvishi, R. Azmi, Fatemeh Moradian, Maral Zarvani","doi":"10.1109/CSICC58665.2023.10105392","DOIUrl":null,"url":null,"abstract":"Owing to the rising standard of living, personal appearance, mainly matching clothes, is essential to people. Because the right clothes not only can beautify their appearance directly but can also increase their self-confidence. This study aims to help users find a matching pair of clothes by considering the intricate details to choose suitable and compatible clothes. In this paper, increasing the efficiency of feature extraction is very important because fashion has a complicated concept, and the extraction of features such as the overall shape, design, and texture of clothes can significantly impact understanding and learning the compatibility of clothes. Therefore, suitable global features can help a lot in understanding the compatibility of clothes. Transformers can extract global features better than convolution networks. We use Swin Transformer networks to extract the image features. We have trained a Triplet-Swin network to learn fashion compatibility, which achieves better accuracy than previous methods. We evaluated our model with AUC and FITB metrics and the Polyvore Outfit dataset.","PeriodicalId":127277,"journal":{"name":"2023 28th International Computer Conference, Computer Society of Iran (CSICC)","volume":"27 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-01-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 28th International Computer Conference, Computer Society of Iran (CSICC)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CSICC58665.2023.10105392","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Owing to the rising standard of living, personal appearance, mainly matching clothes, is essential to people. Because the right clothes not only can beautify their appearance directly but can also increase their self-confidence. This study aims to help users find a matching pair of clothes by considering the intricate details to choose suitable and compatible clothes. In this paper, increasing the efficiency of feature extraction is very important because fashion has a complicated concept, and the extraction of features such as the overall shape, design, and texture of clothes can significantly impact understanding and learning the compatibility of clothes. Therefore, suitable global features can help a lot in understanding the compatibility of clothes. Transformers can extract global features better than convolution networks. We use Swin Transformer networks to extract the image features. We have trained a Triplet-Swin network to learn fashion compatibility, which achieves better accuracy than previous methods. We evaluated our model with AUC and FITB metrics and the Polyvore Outfit dataset.