Rafael Alfaro-Flores, José Salas-Bonilla, Loic Juillard, Juan Esquivel-Rodríguez
{"title":"通过基于重要性的A/B测试,实验驱动的人类在循环机器学习注释的改进","authors":"Rafael Alfaro-Flores, José Salas-Bonilla, Loic Juillard, Juan Esquivel-Rodríguez","doi":"10.1109/CLEI53233.2021.9639977","DOIUrl":null,"url":null,"abstract":"We present an end-to-end experimentation framework to improve the human annotation of data sets used in the training process of Machine Learning models. It covers the instrumentation of the annotation tool, the aggregation of metrics that highlight usage patterns and hypothesis-testing tools that enable the comparison of experimental groups, to decide whether improvements in the annotation process significantly impact the overall results. We show the potential of the protocol using two real-life annotation use cases.","PeriodicalId":6803,"journal":{"name":"2021 XLVII Latin American Computing Conference (CLEI)","volume":"74 1","pages":"1-9"},"PeriodicalIF":0.0000,"publicationDate":"2021-10-25","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Experiment-driven improvements in Human-in-the-loop Machine Learning Annotation via significance-based A/B testing\",\"authors\":\"Rafael Alfaro-Flores, José Salas-Bonilla, Loic Juillard, Juan Esquivel-Rodríguez\",\"doi\":\"10.1109/CLEI53233.2021.9639977\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"We present an end-to-end experimentation framework to improve the human annotation of data sets used in the training process of Machine Learning models. It covers the instrumentation of the annotation tool, the aggregation of metrics that highlight usage patterns and hypothesis-testing tools that enable the comparison of experimental groups, to decide whether improvements in the annotation process significantly impact the overall results. We show the potential of the protocol using two real-life annotation use cases.\",\"PeriodicalId\":6803,\"journal\":{\"name\":\"2021 XLVII Latin American Computing Conference (CLEI)\",\"volume\":\"74 1\",\"pages\":\"1-9\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-10-25\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"2021 XLVII Latin American Computing Conference (CLEI)\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.1109/CLEI53233.2021.9639977\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 XLVII Latin American Computing Conference (CLEI)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/CLEI53233.2021.9639977","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Experiment-driven improvements in Human-in-the-loop Machine Learning Annotation via significance-based A/B testing
We present an end-to-end experimentation framework to improve the human annotation of data sets used in the training process of Machine Learning models. It covers the instrumentation of the annotation tool, the aggregation of metrics that highlight usage patterns and hypothesis-testing tools that enable the comparison of experimental groups, to decide whether improvements in the annotation process significantly impact the overall results. We show the potential of the protocol using two real-life annotation use cases.