{"title":"监督跨模态检索的自适应余量排序","authors":"Tianyuan Xu, Xueliang Liu","doi":"10.1145/3507548.3507599","DOIUrl":null,"url":null,"abstract":"Cross-modal retrieval is to achieve flexible query between different modalities. Many approaches solve the problem by learning a common feature space under to separate the multimodal instances from different categories. But it is challenge to design an effective projecting function. In this paper, we propose a novel cross-modal retrieval method, called Adaptive Margin Ranking for Supervised Cross-modal Retrieval (AMRS). In the solution, we design a neural network as the nonlinear mapping function. To maximize the discrimination of multimodal feature in common representation space, we keep away the samples with different semantic by an adaptive margin, and jointly force the modality invariance to eliminate cross-modal discrepancy. Experimental results on widely used benchmark datasets show that the proposed method is effective in cross-modal learning.","PeriodicalId":414908,"journal":{"name":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","volume":"55 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-12-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Adaptive Margin Ranking for Supervised Cross-modal Retrieval\",\"authors\":\"Tianyuan Xu, Xueliang Liu\",\"doi\":\"10.1145/3507548.3507599\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Cross-modal retrieval is to achieve flexible query between different modalities. Many approaches solve the problem by learning a common feature space under to separate the multimodal instances from different categories. But it is challenge to design an effective projecting function. In this paper, we propose a novel cross-modal retrieval method, called Adaptive Margin Ranking for Supervised Cross-modal Retrieval (AMRS). In the solution, we design a neural network as the nonlinear mapping function. To maximize the discrimination of multimodal feature in common representation space, we keep away the samples with different semantic by an adaptive margin, and jointly force the modality invariance to eliminate cross-modal discrepancy. Experimental results on widely used benchmark datasets show that the proposed method is effective in cross-modal learning.\",\"PeriodicalId\":414908,\"journal\":{\"name\":\"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence\",\"volume\":\"55 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-12-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3507548.3507599\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2021 5th International Conference on Computer Science and Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3507548.3507599","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Adaptive Margin Ranking for Supervised Cross-modal Retrieval
Cross-modal retrieval is to achieve flexible query between different modalities. Many approaches solve the problem by learning a common feature space under to separate the multimodal instances from different categories. But it is challenge to design an effective projecting function. In this paper, we propose a novel cross-modal retrieval method, called Adaptive Margin Ranking for Supervised Cross-modal Retrieval (AMRS). In the solution, we design a neural network as the nonlinear mapping function. To maximize the discrimination of multimodal feature in common representation space, we keep away the samples with different semantic by an adaptive margin, and jointly force the modality invariance to eliminate cross-modal discrepancy. Experimental results on widely used benchmark datasets show that the proposed method is effective in cross-modal learning.