Lovre Torbarina , Tin Ferkovic , Lukasz Roguski , Velimir Mihelcic, Bruno Sarlija, Zeljko Kraljevic
{"title":"通过 ML 生命周期在 NLP 中使用基于变压器的多任务学习的挑战与机遇:立场文件","authors":"Lovre Torbarina , Tin Ferkovic , Lukasz Roguski , Velimir Mihelcic, Bruno Sarlija, Zeljko Kraljevic","doi":"10.1016/j.nlp.2024.100076","DOIUrl":null,"url":null,"abstract":"<div><p>The increasing adoption of natural language processing (NLP) models across industries has led to practitioners’ need for machine learning (ML) systems to handle these models efficiently, from training to serving them in production. However, training, deploying, and updating multiple models can be complex, costly, and time-consuming, mainly when using transformer-based pre-trained language models. Multi-Task Learning (MTL) has emerged as a promising approach to improve efficiency and performance through joint training, rather than training separate models. Motivated by this, we present an overview of MTL approaches in NLP, followed by an in-depth discussion of our position on opportunities they introduce to a set of challenges across various ML lifecycle phases including data engineering, model development, deployment, and monitoring. Our position emphasizes the role of transformer-based MTL approaches in streamlining these lifecycle phases, and we assert that our systematic analysis demonstrates how transformer-based MTL in NLP effectively integrates into ML lifecycle phases. Furthermore, we hypothesize that developing a model that combines MTL for periodic re-training, and continual learning for continual updates and new capabilities integration could be practical, although its viability and effectiveness still demand a substantial empirical investigation.</p></div>","PeriodicalId":100944,"journal":{"name":"Natural Language Processing Journal","volume":"7 ","pages":"Article 100076"},"PeriodicalIF":0.0000,"publicationDate":"2024-04-30","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"https://www.sciencedirect.com/science/article/pii/S2949719124000244/pdfft?md5=9be47fda7d1ff816f43310f77a7417c3&pid=1-s2.0-S2949719124000244-main.pdf","citationCount":"0","resultStr":"{\"title\":\"Challenges and Opportunities of Using Transformer-Based Multi-Task Learning in NLP Through ML Lifecycle: A Position Paper\",\"authors\":\"Lovre Torbarina , Tin Ferkovic , Lukasz Roguski , Velimir Mihelcic, Bruno Sarlija, Zeljko Kraljevic\",\"doi\":\"10.1016/j.nlp.2024.100076\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"<div><p>The increasing adoption of natural language processing (NLP) models across industries has led to practitioners’ need for machine learning (ML) systems to handle these models efficiently, from training to serving them in production. However, training, deploying, and updating multiple models can be complex, costly, and time-consuming, mainly when using transformer-based pre-trained language models. Multi-Task Learning (MTL) has emerged as a promising approach to improve efficiency and performance through joint training, rather than training separate models. Motivated by this, we present an overview of MTL approaches in NLP, followed by an in-depth discussion of our position on opportunities they introduce to a set of challenges across various ML lifecycle phases including data engineering, model development, deployment, and monitoring. Our position emphasizes the role of transformer-based MTL approaches in streamlining these lifecycle phases, and we assert that our systematic analysis demonstrates how transformer-based MTL in NLP effectively integrates into ML lifecycle phases. Furthermore, we hypothesize that developing a model that combines MTL for periodic re-training, and continual learning for continual updates and new capabilities integration could be practical, although its viability and effectiveness still demand a substantial empirical investigation.</p></div>\",\"PeriodicalId\":100944,\"journal\":{\"name\":\"Natural Language Processing Journal\",\"volume\":\"7 \",\"pages\":\"Article 100076\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2024-04-30\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"https://www.sciencedirect.com/science/article/pii/S2949719124000244/pdfft?md5=9be47fda7d1ff816f43310f77a7417c3&pid=1-s2.0-S2949719124000244-main.pdf\",\"citationCount\":\"0\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"Natural Language Processing Journal\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://www.sciencedirect.com/science/article/pii/S2949719124000244\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"Natural Language Processing Journal","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2949719124000244","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0
摘要
随着各行各业越来越多地采用自然语言处理(NLP)模型,从业人员需要机器学习(ML)系统来高效处理这些模型,从训练到为生产提供服务。然而,主要在使用基于转换器的预训练语言模型时,训练、部署和更新多个模型可能会非常复杂、昂贵和耗时。多任务学习(Multi-Task Learning,MTL)是一种很有前途的方法,它可以通过联合训练而不是单独训练模型来提高效率和性能。受此启发,我们概述了 NLP 中的 MTL 方法,然后深入讨论了我们的立场,即这些方法为包括数据工程、模型开发、部署和监控在内的各个 ML 生命周期阶段的一系列挑战带来了哪些机遇。我们的立场强调了基于变压器的 MTL 方法在简化这些生命周期阶段中的作用,并断言我们的系统分析证明了 NLP 中基于变压器的 MTL 如何有效地融入 ML 生命周期各阶段。此外,我们还假设,开发一种将用于定期再培训的 MTL 与用于持续更新和新能力集成的持续学习相结合的模型是切实可行的,尽管其可行性和有效性仍需要大量的实证调查。
Challenges and Opportunities of Using Transformer-Based Multi-Task Learning in NLP Through ML Lifecycle: A Position Paper
The increasing adoption of natural language processing (NLP) models across industries has led to practitioners’ need for machine learning (ML) systems to handle these models efficiently, from training to serving them in production. However, training, deploying, and updating multiple models can be complex, costly, and time-consuming, mainly when using transformer-based pre-trained language models. Multi-Task Learning (MTL) has emerged as a promising approach to improve efficiency and performance through joint training, rather than training separate models. Motivated by this, we present an overview of MTL approaches in NLP, followed by an in-depth discussion of our position on opportunities they introduce to a set of challenges across various ML lifecycle phases including data engineering, model development, deployment, and monitoring. Our position emphasizes the role of transformer-based MTL approaches in streamlining these lifecycle phases, and we assert that our systematic analysis demonstrates how transformer-based MTL in NLP effectively integrates into ML lifecycle phases. Furthermore, we hypothesize that developing a model that combines MTL for periodic re-training, and continual learning for continual updates and new capabilities integration could be practical, although its viability and effectiveness still demand a substantial empirical investigation.