Felix Schneider, Sebastian Stüker, V. Parthasarathy
{"title":"团队缩放@ AutoMin 2021:自动记录的跨域预训练","authors":"Felix Schneider, Sebastian Stüker, V. Parthasarathy","doi":"10.21437/automin.2021-11","DOIUrl":null,"url":null,"abstract":"This Paper describes Zoom’s submission to the First Shared Task on Automatic Minuting at Interspeech 2021. We participated in Task A: generating abstractive summaries of meetings. For this task, we use a transformer-based summarization model which is first trained on data from a similar domain and then finetuned for domain transfer. In this configuration, our model does not yet produce usable summaries. We theorize that in the choice of pretraining corpus, the target side is more important than the source.","PeriodicalId":186820,"journal":{"name":"First Shared Task on Automatic Minuting at Interspeech 2021","volume":"4 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2021-09-04","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"1","resultStr":"{\"title\":\"Team Zoom @ AutoMin 2021: Cross-domain Pretraining for Automatic Minuting\",\"authors\":\"Felix Schneider, Sebastian Stüker, V. Parthasarathy\",\"doi\":\"10.21437/automin.2021-11\",\"DOIUrl\":null,\"url\":null,\"abstract\":\"This Paper describes Zoom’s submission to the First Shared Task on Automatic Minuting at Interspeech 2021. We participated in Task A: generating abstractive summaries of meetings. For this task, we use a transformer-based summarization model which is first trained on data from a similar domain and then finetuned for domain transfer. In this configuration, our model does not yet produce usable summaries. We theorize that in the choice of pretraining corpus, the target side is more important than the source.\",\"PeriodicalId\":186820,\"journal\":{\"name\":\"First Shared Task on Automatic Minuting at Interspeech 2021\",\"volume\":\"4 1\",\"pages\":\"0\"},\"PeriodicalIF\":0.0000,\"publicationDate\":\"2021-09-04\",\"publicationTypes\":\"Journal Article\",\"fieldsOfStudy\":null,\"isOpenAccess\":false,\"openAccessPdf\":\"\",\"citationCount\":\"1\",\"resultStr\":null,\"platform\":\"Semanticscholar\",\"paperid\":null,\"PeriodicalName\":\"First Shared Task on Automatic Minuting at Interspeech 2021\",\"FirstCategoryId\":\"1085\",\"ListUrlMain\":\"https://doi.org/10.21437/automin.2021-11\",\"RegionNum\":0,\"RegionCategory\":null,\"ArticlePicture\":[],\"TitleCN\":null,\"AbstractTextCN\":null,\"PMCID\":null,\"EPubDate\":\"\",\"PubModel\":\"\",\"JCR\":\"\",\"JCRName\":\"\",\"Score\":null,\"Total\":0}","platform":"Semanticscholar","paperid":null,"PeriodicalName":"First Shared Task on Automatic Minuting at Interspeech 2021","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.21437/automin.2021-11","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
Team Zoom @ AutoMin 2021: Cross-domain Pretraining for Automatic Minuting
This Paper describes Zoom’s submission to the First Shared Task on Automatic Minuting at Interspeech 2021. We participated in Task A: generating abstractive summaries of meetings. For this task, we use a transformer-based summarization model which is first trained on data from a similar domain and then finetuned for domain transfer. In this configuration, our model does not yet produce usable summaries. We theorize that in the choice of pretraining corpus, the target side is more important than the source.