Sandy Chen, Leqi Zeng, Abhinav Raghunathan, Flora Huang, Terrence C. Kim
{"title":"MoA is All You Need: Building LLM Research Team using Mixture of Agents","authors":"Sandy Chen, Leqi Zeng, Abhinav Raghunathan, Flora Huang, Terrence C. Kim","doi":"arxiv-2409.07487","DOIUrl":null,"url":null,"abstract":"Large Language Models (LLMs) research in the financial domain is particularly\ncomplex due to the sheer number of approaches proposed in literature.\nRetrieval-Augmented Generation (RAG) has emerged as one of the leading methods\nin the sector due to its inherent groundedness and data source variability. In\nthis work, we introduce a RAG framework called Mixture of Agents (MoA) and\ndemonstrate its viability as a practical, customizable, and highly effective\napproach for scaling RAG applications. MoA is essentially a layered network of\nindividually customized small language models (Hoffmann et al., 2022)\ncollaborating to answer questions and extract information. While there are many\ntheoretical propositions for such an architecture and even a few libraries for\ngenerally applying the structure in practice, there are limited documented\nstudies evaluating the potential of this framework considering real business\nconstraints such as cost and speed. We find that the MoA framework, consisting\nof small language models (Hoffmann et al., 2022), produces higher quality and\nmore grounded responses across various financial domains that are core to\nVanguard's business while simultaneously maintaining low costs.","PeriodicalId":501294,"journal":{"name":"arXiv - QuantFin - Computational Finance","volume":"27 1","pages":""},"PeriodicalIF":0.0000,"publicationDate":"2024-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - QuantFin - Computational Finance","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.07487","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
Large Language Models (LLMs) research in the financial domain is particularly
complex due to the sheer number of approaches proposed in literature.
Retrieval-Augmented Generation (RAG) has emerged as one of the leading methods
in the sector due to its inherent groundedness and data source variability. In
this work, we introduce a RAG framework called Mixture of Agents (MoA) and
demonstrate its viability as a practical, customizable, and highly effective
approach for scaling RAG applications. MoA is essentially a layered network of
individually customized small language models (Hoffmann et al., 2022)
collaborating to answer questions and extract information. While there are many
theoretical propositions for such an architecture and even a few libraries for
generally applying the structure in practice, there are limited documented
studies evaluating the potential of this framework considering real business
constraints such as cost and speed. We find that the MoA framework, consisting
of small language models (Hoffmann et al., 2022), produces higher quality and
more grounded responses across various financial domains that are core to
Vanguard's business while simultaneously maintaining low costs.