Yang Xiao, Yunke Li, Shaoyujie Chen, Hayden Barker, Ryan Rad
{"title":"你真的需要法学硕士学位吗?重新思考客户评论分析的语言模型","authors":"Yang Xiao, Yunke Li, Shaoyujie Chen, Hayden Barker, Ryan Rad","doi":"10.1007/s10462-025-11308-5","DOIUrl":null,"url":null,"abstract":"<div><p>LLarge language models (LLMs) demonstrate strong natural language processing capabilities but come with significant computational costs, raising questions about their practical utility compared to small language models (SLMs). This study systematically compares SLMs (DistilBERT, ELECTRA) and LLMs (Flan-T5, Flan-UL2) on two customer review analysis tasks: sentiment polarity classification and product correlation analysis. Our results show that while LLMs outperform in sentiment classification, they do so at a much higher computational cost, whereas fine-tuned SLMs excel in domain-specific correlation analysis with greater efficiency. To balance accuracy and cost, we propose a context-enhanced hybrid (CE-Hybrid) model, which refines traditional hybrid methods by enriching LLM input with SLM-generated insights, reducing redundant computation while maintaining accuracy. Our findings quantify the trade-offs between model performance and resource efficiency, offering actionable insights for businesses to optimize AI deployment. These results have significant implications for real-world applications such as e-commerce, customer service automation, and business analytics.</p></div>","PeriodicalId":8449,"journal":{"name":"Artificial Intelligence Review","volume":"58 10","pages":""},"PeriodicalIF":13.9000,"publicationDate":"2025-08-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://link.springer.com/content/pdf/10.1007/s10462-025-11308-5.pdf","citationCount":"0","resultStr":"{\"title\":\"Do you actually need an LLM? Rethinking language models for customer reviews analysis\",\"authors\":\"Yang Xiao, Yunke Li, Shaoyujie Chen, Hayden Barker, Ryan Rad\",\"doi\":\"10.1007/s10462-025-11308-5\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>LLarge language models (LLMs) demonstrate strong natural language processing capabilities but come with significant computational costs, raising questions about their practical utility compared to small language models (SLMs). This study systematically compares SLMs (DistilBERT, ELECTRA) and LLMs (Flan-T5, Flan-UL2) on two customer review analysis tasks: sentiment polarity classification and product correlation analysis. Our results show that while LLMs outperform in sentiment classification, they do so at a much higher computational cost, whereas fine-tuned SLMs excel in domain-specific correlation analysis with greater efficiency. To balance accuracy and cost, we propose a context-enhanced hybrid (CE-Hybrid) model, which refines traditional hybrid methods by enriching LLM input with SLM-generated insights, reducing redundant computation while maintaining accuracy. Our findings quantify the trade-offs between model performance and resource efficiency, offering actionable insights for businesses to optimize AI deployment. These results have significant implications for real-world applications such as e-commerce, customer service automation, and business analytics.</p></div>\",\"PeriodicalId\":8449,\"journal\":{\"name\":\"Artificial Intelligence Review\",\"volume\":\"58 10\",\"pages\":\"\"},\"PeriodicalIF\":13.9000,\"publicationDate\":\"2025-08-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://link.springer.com/content/pdf/10.1007/s10462-025-11308-5.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Artificial Intelligence Review\",\"FirstCategoryId\":\"94\",\"ListUrlMain\":\"https://link.springer.com/article/10.1007/s10462-025-11308-5\",\"RegionNum\":2,\"RegionCategory\":\"计算机科学\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Artificial Intelligence Review","FirstCategoryId":"94","ListUrlMain":"https://link.springer.com/article/10.1007/s10462-025-11308-5","RegionNum":2,"RegionCategory":"计算机科学","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
Do you actually need an LLM? Rethinking language models for customer reviews analysis
LLarge language models (LLMs) demonstrate strong natural language processing capabilities but come with significant computational costs, raising questions about their practical utility compared to small language models (SLMs). This study systematically compares SLMs (DistilBERT, ELECTRA) and LLMs (Flan-T5, Flan-UL2) on two customer review analysis tasks: sentiment polarity classification and product correlation analysis. Our results show that while LLMs outperform in sentiment classification, they do so at a much higher computational cost, whereas fine-tuned SLMs excel in domain-specific correlation analysis with greater efficiency. To balance accuracy and cost, we propose a context-enhanced hybrid (CE-Hybrid) model, which refines traditional hybrid methods by enriching LLM input with SLM-generated insights, reducing redundant computation while maintaining accuracy. Our findings quantify the trade-offs between model performance and resource efficiency, offering actionable insights for businesses to optimize AI deployment. These results have significant implications for real-world applications such as e-commerce, customer service automation, and business analytics.
期刊介绍:
Artificial Intelligence Review, a fully open access journal, publishes cutting-edge research in artificial intelligence and cognitive science. It features critical evaluations of applications, techniques, and algorithms, providing a platform for both researchers and application developers. The journal includes refereed survey and tutorial articles, along with reviews and commentary on significant developments in the field.