{"title":"[大型语言模型的技术基础]。","authors":"Christian Blüthgen","doi":"10.1007/s00117-025-01427-z","DOIUrl":null,"url":null,"abstract":"<p><strong>Background: </strong>Large language models (LLMs) such as ChatGPT have rapidly revolutionized the way computers can analyze human language and the way we can interact with computers.</p><p><strong>Objective: </strong>To give an overview of the emergence and basic principles of computational language models.</p><p><strong>Methods: </strong>Narrative literature-based analysis of the history of the emergence of language models, the technical foundations, the training process and the limitations of LLMs.</p><p><strong>Results: </strong>Nowadays, LLMs are mostly based on transformer models that can capture context through their attention mechanism. Through a multistage training process with comprehensive pretraining, supervised fine-tuning and alignment with human preferences, LLMs have developed a general understanding of language. This enables them to flexibly analyze texts and produce outputs of high linguistic quality.</p><p><strong>Conclusion: </strong>Their technical foundations and training process make large language models versatile general-purpose tools for text processing, with numerous applications in radiology. The main limitation is the tendency to postulate incorrect but plausible-sounding information with high confidence.</p>","PeriodicalId":74635,"journal":{"name":"Radiologie (Heidelberg, Germany)","volume":" ","pages":"227-234"},"PeriodicalIF":0.6000,"publicationDate":"2025-04-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11937190/pdf/","citationCount":"0","resultStr":"{\"title\":\"[Technical foundations of large language models].\",\"authors\":\"Christian Blüthgen\",\"doi\":\"10.1007/s00117-025-01427-z\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<p><strong>Background: </strong>Large language models (LLMs) such as ChatGPT have rapidly revolutionized the way computers can analyze human language and the way we can interact with computers.</p><p><strong>Objective: </strong>To give an overview of the emergence and basic principles of computational language models.</p><p><strong>Methods: </strong>Narrative literature-based analysis of the history of the emergence of language models, the technical foundations, the training process and the limitations of LLMs.</p><p><strong>Results: </strong>Nowadays, LLMs are mostly based on transformer models that can capture context through their attention mechanism. Through a multistage training process with comprehensive pretraining, supervised fine-tuning and alignment with human preferences, LLMs have developed a general understanding of language. This enables them to flexibly analyze texts and produce outputs of high linguistic quality.</p><p><strong>Conclusion: </strong>Their technical foundations and training process make large language models versatile general-purpose tools for text processing, with numerous applications in radiology. The main limitation is the tendency to postulate incorrect but plausible-sounding information with high confidence.</p>\",\"PeriodicalId\":74635,\"journal\":{\"name\":\"Radiologie (Heidelberg, Germany)\",\"volume\":\" \",\"pages\":\"227-234\"},\"PeriodicalIF\":0.6000,\"publicationDate\":\"2025-04-01\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.ncbi.nlm.nih.gov/pmc/articles/PMC11937190/pdf/\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Radiologie (Heidelberg, Germany)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1007/s00117-025-01427-z\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"2025/3/10 0:00:00\",\"PubModel\":\"Epub\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Radiologie (Heidelberg, Germany)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1007/s00117-025-01427-z","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2025/3/10 0:00:00","PubModel":"Epub","JCR":"","JCRName":"","Score":null,"Total":0}
Background: Large language models (LLMs) such as ChatGPT have rapidly revolutionized the way computers can analyze human language and the way we can interact with computers.
Objective: To give an overview of the emergence and basic principles of computational language models.
Methods: Narrative literature-based analysis of the history of the emergence of language models, the technical foundations, the training process and the limitations of LLMs.
Results: Nowadays, LLMs are mostly based on transformer models that can capture context through their attention mechanism. Through a multistage training process with comprehensive pretraining, supervised fine-tuning and alignment with human preferences, LLMs have developed a general understanding of language. This enables them to flexibly analyze texts and produce outputs of high linguistic quality.
Conclusion: Their technical foundations and training process make large language models versatile general-purpose tools for text processing, with numerous applications in radiology. The main limitation is the tendency to postulate incorrect but plausible-sounding information with high confidence.