Zhuorui Zhang , Shanshan Feng , Tiance Yang , Ruobing Huang , Hao Wang , Fu Wang , Fan Li
{"title":"AviationCopilot:受人类飞行员训练启发,建立一个可靠的基于llm的航空副驾驶","authors":"Zhuorui Zhang , Shanshan Feng , Tiance Yang , Ruobing Huang , Hao Wang , Fu Wang , Fan Li","doi":"10.1016/j.aei.2025.103806","DOIUrl":null,"url":null,"abstract":"<div><div>Modern pilots routinely face high cognitive loads during complex flight operations. Although large language models (LLMs) demonstrate exceptional natural language understanding and exhibit tremendous potential as copilots, there is a notable gap in LLMs specifically designed to handle the knowledge-intensive tasks required of pilots. Inspired by pilots’ learning and manual retrieval patterns, we introduce <strong>AviationCopilot</strong>—a novel framework that efficiently injects both aviation knowledge content and knowledge structure into LLMs. Specifically, we employ differentiated data fusion and generalization strategies for two training stages including continual pre-training and instruction tuning. This approach equips the model with enhanced domain-specific knowledge retention and instruction-following capabilities, akin to human pilots. During inference, AviationCopilot activates its knowledge structure memory to adaptively retrieve comprehensive context, improving factual accuracy. To evaluate effectiveness, we construct a comprehensive benchmark named <strong>OpenAviation</strong> featuring both LLM-synthesized and expert-designed questions. Experimental results show that models with fewer than two billion parameters, trained with the AviationCopilot framework, consistently outperform strong LLM baselines, including those utilizing Retrieval-Augmented Generation (RAG). Additionally, AviationCopilot enhances structured aviation understanding and enables LLMs to serve as retrievers for improving other models, supporting more reliable AI copilots. The dataset and source code are available at <span><span>https://github.com/zhuorui-zhang/AviationCopilot</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":50941,"journal":{"name":"Advanced Engineering Informatics","volume":"69 ","pages":"Article 103806"},"PeriodicalIF":9.9000,"publicationDate":"2025-09-08","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":"{\"title\":\"AviationCopilot: Building a reliable LLM-based Aviation Copilot inspired by human pilot training\",\"authors\":\"Zhuorui Zhang , Shanshan Feng , Tiance Yang , Ruobing Huang , Hao Wang , Fu Wang , Fan Li\",\"doi\":\"10.1016/j.aei.2025.103806\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><div>Modern pilots routinely face high cognitive loads during complex flight operations. Although large language models (LLMs) demonstrate exceptional natural language understanding and exhibit tremendous potential as copilots, there is a notable gap in LLMs specifically designed to handle the knowledge-intensive tasks required of pilots. Inspired by pilots’ learning and manual retrieval patterns, we introduce <strong>AviationCopilot</strong>—a novel framework that efficiently injects both aviation knowledge content and knowledge structure into LLMs. Specifically, we employ differentiated data fusion and generalization strategies for two training stages including continual pre-training and instruction tuning. This approach equips the model with enhanced domain-specific knowledge retention and instruction-following capabilities, akin to human pilots. During inference, AviationCopilot activates its knowledge structure memory to adaptively retrieve comprehensive context, improving factual accuracy. To evaluate effectiveness, we construct a comprehensive benchmark named <strong>OpenAviation</strong> featuring both LLM-synthesized and expert-designed questions. Experimental results show that models with fewer than two billion parameters, trained with the AviationCopilot framework, consistently outperform strong LLM baselines, including those utilizing Retrieval-Augmented Generation (RAG). Additionally, AviationCopilot enhances structured aviation understanding and enables LLMs to serve as retrievers for improving other models, supporting more reliable AI copilots. The dataset and source code are available at <span><span>https://github.com/zhuorui-zhang/AviationCopilot</span><svg><path></path></svg></span>.</div></div>\",\"PeriodicalId\":50941,\"journal\":{\"name\":\"Advanced Engineering Informatics\",\"volume\":\"69 \",\"pages\":\"Article 103806\"},\"PeriodicalIF\":9.9000,\"publicationDate\":\"2025-09-08\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Advanced Engineering Informatics\",\"FirstCategoryId\":\"5\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S1474034625006998\",\"RegionNum\":1,\"RegionCategory\":\"工程技术\",\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"Q1\",\"JCRName\":\"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Advanced Engineering Informatics","FirstCategoryId":"5","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S1474034625006998","RegionNum":1,"RegionCategory":"工程技术","ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"Q1","JCRName":"COMPUTER SCIENCE, ARTIFICIAL INTELLIGENCE","Score":null,"Total":0}
AviationCopilot: Building a reliable LLM-based Aviation Copilot inspired by human pilot training
Modern pilots routinely face high cognitive loads during complex flight operations. Although large language models (LLMs) demonstrate exceptional natural language understanding and exhibit tremendous potential as copilots, there is a notable gap in LLMs specifically designed to handle the knowledge-intensive tasks required of pilots. Inspired by pilots’ learning and manual retrieval patterns, we introduce AviationCopilot—a novel framework that efficiently injects both aviation knowledge content and knowledge structure into LLMs. Specifically, we employ differentiated data fusion and generalization strategies for two training stages including continual pre-training and instruction tuning. This approach equips the model with enhanced domain-specific knowledge retention and instruction-following capabilities, akin to human pilots. During inference, AviationCopilot activates its knowledge structure memory to adaptively retrieve comprehensive context, improving factual accuracy. To evaluate effectiveness, we construct a comprehensive benchmark named OpenAviation featuring both LLM-synthesized and expert-designed questions. Experimental results show that models with fewer than two billion parameters, trained with the AviationCopilot framework, consistently outperform strong LLM baselines, including those utilizing Retrieval-Augmented Generation (RAG). Additionally, AviationCopilot enhances structured aviation understanding and enables LLMs to serve as retrievers for improving other models, supporting more reliable AI copilots. The dataset and source code are available at https://github.com/zhuorui-zhang/AviationCopilot.
期刊介绍:
Advanced Engineering Informatics is an international Journal that solicits research papers with an emphasis on 'knowledge' and 'engineering applications'. The Journal seeks original papers that report progress in applying methods of engineering informatics. These papers should have engineering relevance and help provide a scientific base for more reliable, spontaneous, and creative engineering decision-making. Additionally, papers should demonstrate the science of supporting knowledge-intensive engineering tasks and validate the generality, power, and scalability of new methods through rigorous evaluation, preferably both qualitatively and quantitatively. Abstracting and indexing for Advanced Engineering Informatics include Science Citation Index Expanded, Scopus and INSPEC.