{"title":"情感分析的文本表示:从静态到动态","authors":"P. M. Gavali, Suresh K. Shiragave","doi":"10.1109/ICSMDI57622.2023.00025","DOIUrl":null,"url":null,"abstract":"Text representation in a vector, known as embedding, is crucial for various classification tasks including sentiment analysis. It helps to process and understand natural language text more effectively. It has evolved from static approaches, such as bag-of-words and n-grams, to more dynamic approaches that consider the context and meaning of words, such as word embeddings and contextualized embeddings. Word embeddings use neural networks to learn vector representations of words based on their co-occurrence patterns in large text corpora. On the other hand, contextualized embeddings, such as BERT, consider the context of each word within a sentence or document to generate more nuanced representations. Numerous researchers have suggested modifying the original Word2Vec and BERT embeddings to include sentiment information. This paper provides a comprehensive overview of these methods by including a detailed discussion of various evaluation techniques. The paper also outlines several challenges related to embeddings that can be addressed in order to improve the results of sentiment analysis.","PeriodicalId":373017,"journal":{"name":"2023 3rd International Conference on Smart Data Intelligence (ICSMDI)","volume":"107 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2023-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"Text Representation for Sentiment Analysis: From Static to Dynamic\",\"authors\":\"P. M. Gavali, Suresh K. Shiragave\",\"doi\":\"10.1109/ICSMDI57622.2023.00025\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"Text representation in a vector, known as embedding, is crucial for various classification tasks including sentiment analysis. It helps to process and understand natural language text more effectively. It has evolved from static approaches, such as bag-of-words and n-grams, to more dynamic approaches that consider the context and meaning of words, such as word embeddings and contextualized embeddings. Word embeddings use neural networks to learn vector representations of words based on their co-occurrence patterns in large text corpora. On the other hand, contextualized embeddings, such as BERT, consider the context of each word within a sentence or document to generate more nuanced representations. Numerous researchers have suggested modifying the original Word2Vec and BERT embeddings to include sentiment information. This paper provides a comprehensive overview of these methods by including a detailed discussion of various evaluation techniques. The paper also outlines several challenges related to embeddings that can be addressed in order to improve the results of sentiment analysis.\",\"PeriodicalId\":373017,\"journal\":{\"name\":\"2023 3rd International Conference on Smart Data Intelligence (ICSMDI)\",\"volume\":\"107 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2023-03-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2023 3rd International Conference on Smart Data Intelligence (ICSMDI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/ICSMDI57622.2023.00025\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2023 3rd International Conference on Smart Data Intelligence (ICSMDI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICSMDI57622.2023.00025","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Text Representation for Sentiment Analysis: From Static to Dynamic
Text representation in a vector, known as embedding, is crucial for various classification tasks including sentiment analysis. It helps to process and understand natural language text more effectively. It has evolved from static approaches, such as bag-of-words and n-grams, to more dynamic approaches that consider the context and meaning of words, such as word embeddings and contextualized embeddings. Word embeddings use neural networks to learn vector representations of words based on their co-occurrence patterns in large text corpora. On the other hand, contextualized embeddings, such as BERT, consider the context of each word within a sentence or document to generate more nuanced representations. Numerous researchers have suggested modifying the original Word2Vec and BERT embeddings to include sentiment information. This paper provides a comprehensive overview of these methods by including a detailed discussion of various evaluation techniques. The paper also outlines several challenges related to embeddings that can be addressed in order to improve the results of sentiment analysis.