{"title":"基于SF-Transformer的神经手语翻译","authors":"Qifang Yin, Wenqi Tao, Xiaolong Liu, Yu Hong","doi":"10.1145/3529466.3529503","DOIUrl":null,"url":null,"abstract":"The popular methods are based on the combination of CNNs and RNNs in the sign language translation. Recently, Transformer has also attracted the attention of researchers and achieved success in this subject. However, researchers usually only focus on the accuracy of their model, while ignoring the practical application value. In this paper, we propose the SF-Transformer, a lightweight model based on Encoder-Decoder architecture for sign language translation, which achieves new state-of-the-art performance on Chinese Sign Language (CSL) dataset. We used 2D/3D convolution blocks of SF-Net and Transformer's Decoders to build our network. Benefiting from fewer parameters and a high level of parallelization, the training and inference speed of our model is faster. We hope that our method can contribute to the practical application of sign language translation on low-computing devices such as mobile phones.","PeriodicalId":375562,"journal":{"name":"Proceedings of the 2022 6th International Conference on Innovation in Artificial Intelligence","volume":"22 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-03-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"3","resultStr":"{\"title\":\"Neural Sign Language Translation with SF-Transformer\",\"authors\":\"Qifang Yin, Wenqi Tao, Xiaolong Liu, Yu Hong\",\"doi\":\"10.1145/3529466.3529503\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"The popular methods are based on the combination of CNNs and RNNs in the sign language translation. Recently, Transformer has also attracted the attention of researchers and achieved success in this subject. However, researchers usually only focus on the accuracy of their model, while ignoring the practical application value. In this paper, we propose the SF-Transformer, a lightweight model based on Encoder-Decoder architecture for sign language translation, which achieves new state-of-the-art performance on Chinese Sign Language (CSL) dataset. We used 2D/3D convolution blocks of SF-Net and Transformer's Decoders to build our network. Benefiting from fewer parameters and a high level of parallelization, the training and inference speed of our model is faster. We hope that our method can contribute to the practical application of sign language translation on low-computing devices such as mobile phones.\",\"PeriodicalId\":375562,\"journal\":{\"name\":\"Proceedings of the 2022 6th International Conference on Innovation in Artificial Intelligence\",\"volume\":\"22 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2022-03-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"3\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 2022 6th International Conference on Innovation in Artificial Intelligence\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1145/3529466.3529503\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 2022 6th International Conference on Innovation in Artificial Intelligence","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1145/3529466.3529503","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Neural Sign Language Translation with SF-Transformer
The popular methods are based on the combination of CNNs and RNNs in the sign language translation. Recently, Transformer has also attracted the attention of researchers and achieved success in this subject. However, researchers usually only focus on the accuracy of their model, while ignoring the practical application value. In this paper, we propose the SF-Transformer, a lightweight model based on Encoder-Decoder architecture for sign language translation, which achieves new state-of-the-art performance on Chinese Sign Language (CSL) dataset. We used 2D/3D convolution blocks of SF-Net and Transformer's Decoders to build our network. Benefiting from fewer parameters and a high level of parallelization, the training and inference speed of our model is faster. We hope that our method can contribute to the practical application of sign language translation on low-computing devices such as mobile phones.