{"title":"Adaptive Margin Ranking for Supervised Cross-modal Retrieval","authors":"Tianyuan Xu, Xueliang Liu","doi":"10.1145/3507548.3507599","DOIUrl":null,"url":null,"abstract":"Cross-modal retrieval is to achieve flexible query between different modalities. Many approaches solve the problem by learning a common feature space under to separate the multimodal instances from different categories. But it is challenge to design an effective projecting function. In this paper, we propose a novel cross-modal retrieval method, called Adaptive Margin Ranking for Supervised Cross-modal Retrieval (AMRS). In the solution, we design a neural network as the nonlinear mapping function. To maximize the discrimination of multimodal feature in common representation space, we keep away the samples with different semantic by an adaptive margin, and jointly force the modality invariance to eliminate cross-modal discrepancy. Experimental results on widely used benchmark datasets show that the proposed method is effective in cross-modal learning.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"55 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3507548.3507599","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Cross-modal retrieval is to achieve flexible query between different modalities. Many approaches solve the problem by learning a common feature space under to separate the multimodal instances from different categories. But it is challenge to design an effective projecting function. In this paper, we propose a novel cross-modal retrieval method, called Adaptive Margin Ranking for Supervised Cross-modal Retrieval (AMRS). In the solution, we design a neural network as the nonlinear mapping function. To maximize the discrimination of multimodal feature in common representation space, we keep away the samples with different semantic by an adaptive margin, and jointly force the modality invariance to eliminate cross-modal discrepancy. Experimental results on widely used benchmark datasets show that the proposed method is effective in cross-modal learning.