{"title":"学习改进的类向量多类问题类型分类","authors":"Tanu Gupta, Ela Kumar","doi":"10.2991/ahis.k.210913.015","DOIUrl":null,"url":null,"abstract":"Recent research in NLP has exploited word embedding to achieve outstanding results in various tasks such as; spam filtering, text classification and summarization and others. Present word embedding algorithms have power to capture semantic and syntactic knowledge about word, but not enough to portray the distinct meaning of polysemy word. Many work has utilized sense embeddings to integrate all possible meaning to word vector, which is computationally expensive. Context embedding is another way out to identify word’s actual meaning, but it is hard to enumerate every context with a small size dataset. This paper has proposed a methodology to generate improved class-specific word vector that enhance the distinctive property of word in a class to tackle light polysemy problem in question classification. The proposed approach is compared with baseline approaches, tested using deep learning models upon TREC, Kaggle and Yahoo questions datasets and respectively attain 93.6%, 91.8% and 89.2% accuracy.","PeriodicalId":417648,"journal":{"name":"Proceedings of the 3rd International Conference on Integrated Intelligent Computing Communication & Security (ICIIC 2021)","volume":"433 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"1900-01-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Learning Improved Class Vector for Multi-Class Question Type Classification\",\"authors\":\"Tanu Gupta, Ela Kumar\",\"doi\":\"10.2991/ahis.k.210913.015\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Recent research in NLP has exploited word embedding to achieve outstanding results in various tasks such as; spam filtering, text classification and summarization and others. Present word embedding algorithms have power to capture semantic and syntactic knowledge about word, but not enough to portray the distinct meaning of polysemy word. Many work has utilized sense embeddings to integrate all possible meaning to word vector, which is computationally expensive. Context embedding is another way out to identify word’s actual meaning, but it is hard to enumerate every context with a small size dataset. This paper has proposed a methodology to generate improved class-specific word vector that enhance the distinctive property of word in a class to tackle light polysemy problem in question classification. The proposed approach is compared with baseline approaches, tested using deep learning models upon TREC, Kaggle and Yahoo questions datasets and respectively attain 93.6%, 91.8% and 89.2% accuracy.\",\"PeriodicalId\":417648,\"journal\":{\"name\":\"Proceedings of the 3rd International Conference on Integrated Intelligent Computing Communication & Security (ICIIC 2021)\",\"volume\":\"433 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"1900-01-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Proceedings of the 3rd International Conference on Integrated Intelligent Computing Communication & Security (ICIIC 2021)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.2991/ahis.k.210913.015\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Proceedings of the 3rd International Conference on Integrated Intelligent Computing Communication & Security (ICIIC 2021)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.2991/ahis.k.210913.015","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Learning Improved Class Vector for Multi-Class Question Type Classification
Recent research in NLP has exploited word embedding to achieve outstanding results in various tasks such as; spam filtering, text classification and summarization and others. Present word embedding algorithms have power to capture semantic and syntactic knowledge about word, but not enough to portray the distinct meaning of polysemy word. Many work has utilized sense embeddings to integrate all possible meaning to word vector, which is computationally expensive. Context embedding is another way out to identify word’s actual meaning, but it is hard to enumerate every context with a small size dataset. This paper has proposed a methodology to generate improved class-specific word vector that enhance the distinctive property of word in a class to tackle light polysemy problem in question classification. The proposed approach is compared with baseline approaches, tested using deep learning models upon TREC, Kaggle and Yahoo questions datasets and respectively attain 93.6%, 91.8% and 89.2% accuracy.