{"title":"AMGPT:用于增材制造语境查询的大型语言模型","authors":"","doi":"10.1016/j.addlet.2024.100232","DOIUrl":null,"url":null,"abstract":"<div><p>Generalized large language models (LLMs) such as GPT-4 may not provide specific answers to queries formulated by materials science researchers. These models may produce a high-level outline but lack the capacity to return detailed instructions on manufacturing and material properties of novel alloys. We introduce “AMGPT”, a specialized LLM text generator designed for metal AM queries. The goal of AMGPT is to assist researchers and users in navigating a curated corpus of literature. Instead of training from scratch, we employ a pre-trained Llama2-7B model from Hugging Face in a Retrieval-Augmented Generation (RAG) setup, utilizing it to dynamically incorporate information from <span><math><mo>∼</mo></math></span>50 AM papers and textbooks in PDF format. Mathpix is used to convert these PDF documents into TeX format, facilitating their integration into the RAG pipeline managed by LlamaIndex. A query retrieval function has also been added, enabling the system to fetch relevant literature from Elsevier journals based on the context of the query. Expert evaluations of this project highlight that specific embeddings from the RAG setup accelerate response times and maintain coherence in the generated text.</p></div>","PeriodicalId":72068,"journal":{"name":"Additive manufacturing letters","volume":null,"pages":null},"PeriodicalIF":4.2000,"publicationDate":"2024-08-22","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2772369024000409/pdfft?md5=8d7e38c2365561cad4541597909ff24b&pid=1-s2.0-S2772369024000409-main.pdf","citationCount":"0","resultStr":"{\"title\":\"AMGPT: A large language model for contextual querying in additive manufacturing\",\"authors\":\"\",\"doi\":\"10.1016/j.addlet.2024.100232\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>Generalized large language models (LLMs) such as GPT-4 may not provide specific answers to queries formulated by materials science researchers. These models may produce a high-level outline but lack the capacity to return detailed instructions on manufacturing and material properties of novel alloys. We introduce “AMGPT”, a specialized LLM text generator designed for metal AM queries. The goal of AMGPT is to assist researchers and users in navigating a curated corpus of literature. Instead of training from scratch, we employ a pre-trained Llama2-7B model from Hugging Face in a Retrieval-Augmented Generation (RAG) setup, utilizing it to dynamically incorporate information from <span><math><mo>∼</mo></math></span>50 AM papers and textbooks in PDF format. Mathpix is used to convert these PDF documents into TeX format, facilitating their integration into the RAG pipeline managed by LlamaIndex. A query retrieval function has also been added, enabling the system to fetch relevant literature from Elsevier journals based on the context of the query. Expert evaluations of this project highlight that specific embeddings from the RAG setup accelerate response times and maintain coherence in the generated text.</p></div>\",\"PeriodicalId\":72068,\"journal\":{\"name\":\"Additive manufacturing letters\",\"volume\":null,\"pages\":null},\"PeriodicalIF\":4.2000,\"publicationDate\":\"2024-08-22\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.sciencedirect.com/science/article/pii/S2772369024000409/pdfft?md5=8d7e38c2365561cad4541597909ff24b&pid=1-s2.0-S2772369024000409-main.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Additive manufacturing letters\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2772369024000409\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q2\",\"JCRName\":\"ENGINEERING, MANUFACTURING\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Additive manufacturing letters","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2772369024000409","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q2","JCRName":"ENGINEERING, MANUFACTURING","Score":null,"Total":0}
引用次数: 0
摘要
通用大型语言模型(LLM),如 GPT-4,可能无法为材料科学研究人员提出的查询提供具体答案。这些模型可以生成一个高级大纲,但缺乏返回有关新型合金的制造和材料特性的详细说明的能力。我们介绍了 "AMGPT",这是一种专门为金属 AM 查询设计的 LLM 文本生成器。AMGPT 的目标是帮助研究人员和用户浏览经过整理的文献语料库。我们没有从头开始训练,而是在检索增强生成(RAG)设置中使用了来自 Hugging Face 的预训练 Llama2-7B 模型,并利用它动态纳入了来自 ∼50 篇 AM 论文和 PDF 格式教科书的信息。Mathpix 用于将这些 PDF 文档转换为 TeX 格式,便于将其整合到由 LlamaIndex 管理的 RAG 管道中。系统还增加了查询检索功能,可根据查询内容从爱思唯尔期刊中获取相关文献。该项目的专家评估强调,RAG 设置中的特定嵌入可加快响应时间,并保持生成文本的一致性。
AMGPT: A large language model for contextual querying in additive manufacturing
Generalized large language models (LLMs) such as GPT-4 may not provide specific answers to queries formulated by materials science researchers. These models may produce a high-level outline but lack the capacity to return detailed instructions on manufacturing and material properties of novel alloys. We introduce “AMGPT”, a specialized LLM text generator designed for metal AM queries. The goal of AMGPT is to assist researchers and users in navigating a curated corpus of literature. Instead of training from scratch, we employ a pre-trained Llama2-7B model from Hugging Face in a Retrieval-Augmented Generation (RAG) setup, utilizing it to dynamically incorporate information from 50 AM papers and textbooks in PDF format. Mathpix is used to convert these PDF documents into TeX format, facilitating their integration into the RAG pipeline managed by LlamaIndex. A query retrieval function has also been added, enabling the system to fetch relevant literature from Elsevier journals based on the context of the query. Expert evaluations of this project highlight that specific embeddings from the RAG setup accelerate response times and maintain coherence in the generated text.