C. B. Pronin, A. V. Volosova, A. V. Ostroukh, Yu. N. Strogov
{"title":"Overcoming linguistic barriers in code assistants: creating a QLoRA adapter to improve support for Russian-language code writing instructions","authors":"C. B. Pronin, A. V. Volosova, A. V. Ostroukh, Yu. N. Strogov","doi":"arxiv-2409.09353","DOIUrl":null,"url":null,"abstract":"In this paper, an approach to training and evaluating an adapter model for\nthe popular language model \"zephyr-7b-beta\" is described. The adapter was\ndeveloped to improve the performance of the base model in tasks related to\nprogramming and understanding the Russian language. Considering the high\nquality of the original model in tasks in the English language, the goal of the\nresearch was to expand its linguistic and technical spectrum. The proposed\nadapter was trained using a large and diverse dataset, including\nquestion-answer pairs related to programming, as well code-related texts in\nRussian language. The applied training methodology ensures an improvement in\nthe model's quality of answers in understanding and generating Python code\nbased on Russian instructions. We evaluated the performance of the base model\nwith the installed adapter using various metrics, comparing it to the base\nmodel as well as other state-of-the-art models in this field. The obtained\nresults showed significant improvement, both in tasks related to writing Python\ncode and in processing the Russian language, confirming the effectiveness of\nthe proposed adapter.","PeriodicalId":501278,"journal":{"name":"arXiv - CS - Software Engineering","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2024-09-14","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"arXiv - CS - Software Engineering","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/arxiv-2409.09353","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
Abstract
In this paper, an approach to training and evaluating an adapter model for
the popular language model "zephyr-7b-beta" is described. The adapter was
developed to improve the performance of the base model in tasks related to
programming and understanding the Russian language. Considering the high
quality of the original model in tasks in the English language, the goal of the
research was to expand its linguistic and technical spectrum. The proposed
adapter was trained using a large and diverse dataset, including
question-answer pairs related to programming, as well code-related texts in
Russian language. The applied training methodology ensures an improvement in
the model's quality of answers in understanding and generating Python code
based on Russian instructions. We evaluated the performance of the base model
with the installed adapter using various metrics, comparing it to the base
model as well as other state-of-the-art models in this field. The obtained
results showed significant improvement, both in tasks related to writing Python
code and in processing the Russian language, confirming the effectiveness of
the proposed adapter.