{"title":"The evolution of language models: From N-Grams to LLMs, and beyond","authors":"Mohammad Ghaseminejad Raeini","doi":"10.1016/j.nlp.2025.100168","DOIUrl":null,"url":null,"abstract":"<div><div>In the last couple of decades language models and artificial intelligence technologies have had significant improvements. Along with computer vision and image processing models, large language models (LLMs) are expected to have big impacts on how AI technologies will evolve. As such, it is important to study how language models have advanced since their inception; and more importantly how they will grow in the future.</div><div>In this article, we provide an overview of the evolution of language models. We start with early statistical and rule-based models. The advancement of language models are discussed all the way to nowadays transformer-based multimodal models (MM-LLMs). We discuss the shortcomings of the current language models and various aspects of the models that need to be improved upon. We also highlight the latest research trends in NLP. Furthermore, we pinpoint important aspects of language models and AI technologies that need further attention. This overview paper provides valuable insights about the progression of language models. It can be motivational and helpful for advancing the state-of-art language models.</div></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"12 ","pages":"Article 100168"},"PeriodicalIF":0.0000,"publicationDate":"2025-06-24","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Natural Language Processing Journal","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2949719125000445","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In the last couple of decades language models and artificial intelligence technologies have had significant improvements. Along with computer vision and image processing models, large language models (LLMs) are expected to have big impacts on how AI technologies will evolve. As such, it is important to study how language models have advanced since their inception; and more importantly how they will grow in the future.
In this article, we provide an overview of the evolution of language models. We start with early statistical and rule-based models. The advancement of language models are discussed all the way to nowadays transformer-based multimodal models (MM-LLMs). We discuss the shortcomings of the current language models and various aspects of the models that need to be improved upon. We also highlight the latest research trends in NLP. Furthermore, we pinpoint important aspects of language models and AI technologies that need further attention. This overview paper provides valuable insights about the progression of language models. It can be motivational and helpful for advancing the state-of-art language models.