DisT5: A Text-to-Text transformer model for disaster events

IF 2.9 Q1 Social Sciences
Online Social Networks and Media Pub Date : 2026-03-01 Epub Date: 2026-02-22 DOI:10.1016/j.osnem.2026.100347
Piyush Kumar Garg , Srishti Gupta , Syed Ali Abbas , Roshni Chakraborty , Sourav Kumar Dandapat
{"title":"DisT5: A Text-to-Text transformer model for disaster events","authors":"Piyush Kumar Garg ,&nbsp;Srishti Gupta ,&nbsp;Syed Ali Abbas ,&nbsp;Roshni Chakraborty ,&nbsp;Sourav Kumar Dandapat","doi":"10.1016/j.osnem.2026.100347","DOIUrl":null,"url":null,"abstract":"<div><div>Transformer-based language models have become essential components in constructing pipelines for various NLP tasks. Although various pre-trained transformer models have been developed for different tasks, they still fail to achieve the desired performance in the disaster domain. Consequently, a lack of pre-trained models specifically designed for this domain. Therefore, in this paper, we present DisT5, a pre-trained transformer model specific to the disaster domain. We have adopted a T5-style self-supervised pre-training approach. We further pre-trained the T5-model on a large collection of textual data covering diverse disaster events, both natural and man-made. We benchmark DisT5 on three downstream tasks: (1) Tweet category classification, (2) Key-phrase identification, and (3) Abstractive summarization. We validate the performance of DisT5 for the aforementioned tasks through comprehensive experiments. Our experimental results demonstrate that additional pre-training improves the performance of the DisT5 model in all the aforementioned tasks. We achieved up to 70.35% accuracy improvement in tweet classification, 87.65% IOU (F1-score) improvement in key-phrase identification, and up to 76.47% and 72.72% ROUGE-N F1-score improvement in summarization without and with fine-tuning, respectively. The pre-trained model will be made available at <span><span>https://huggingface.co/Piyyussh/DisT5</span><svg><path></path></svg></span>.</div></div>","PeriodicalId":52228,"journal":{"name":"Online Social Networks and Media","volume":"52 ","pages":"Article 100347"},"PeriodicalIF":2.9000,"publicationDate":"2026-03-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"Online Social Networks and Media","FirstCategoryId":"1085","ListUrlMain":"https://www.sciencedirect.com/science/article/pii/S2468696426000030","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"2026/2/22 0:00:00","PubModel":"Epub","JCR":"Q1","JCRName":"Social Sciences","Score":null,"Total":0}
引用次数: 0

Abstract

Transformer-based language models have become essential components in constructing pipelines for various NLP tasks. Although various pre-trained transformer models have been developed for different tasks, they still fail to achieve the desired performance in the disaster domain. Consequently, a lack of pre-trained models specifically designed for this domain. Therefore, in this paper, we present DisT5, a pre-trained transformer model specific to the disaster domain. We have adopted a T5-style self-supervised pre-training approach. We further pre-trained the T5-model on a large collection of textual data covering diverse disaster events, both natural and man-made. We benchmark DisT5 on three downstream tasks: (1) Tweet category classification, (2) Key-phrase identification, and (3) Abstractive summarization. We validate the performance of DisT5 for the aforementioned tasks through comprehensive experiments. Our experimental results demonstrate that additional pre-training improves the performance of the DisT5 model in all the aforementioned tasks. We achieved up to 70.35% accuracy improvement in tweet classification, 87.65% IOU (F1-score) improvement in key-phrase identification, and up to 76.47% and 72.72% ROUGE-N F1-score improvement in summarization without and with fine-tuning, respectively. The pre-trained model will be made available at https://huggingface.co/Piyyussh/DisT5.
DisT5:灾难事件的文本到文本转换器模型
基于转换器的语言模型已经成为为各种NLP任务构建管道的重要组成部分。尽管针对不同的任务开发了各种预训练的变压器模型,但它们在灾难领域仍然无法达到预期的性能。因此,缺乏专门为该领域设计的预训练模型。因此,在本文中,我们提出了DisT5,一个专门针对灾害领域的预训练变压器模型。我们采用了t5式的自监督预训练方法。我们在涵盖自然和人为灾害事件的大量文本数据上进一步对t5模型进行了预训练。我们在三个下游任务上对DisT5进行基准测试:(1)推文类别分类,(2)关键短语识别,(3)抽象摘要。我们通过综合实验验证了DisT5在上述任务中的性能。我们的实验结果表明,额外的预训练提高了DisT5模型在上述所有任务中的性能。我们在tweet分类上的准确率提高了70.35%,在关键短语识别上的IOU (F1-score)提高了87.65%,在无微调和有微调的总结上的ROUGE-N F1-score分别提高了76.47%和72.72%。预训练模型将在https://huggingface.co/Piyyussh/DisT5上提供。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
Online Social Networks and Media
Online Social Networks and Media Social Sciences-Communication
CiteScore
10.60
自引率
0.00%
发文量
32
审稿时长
44 days
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:604180095
Book学术官方微信
小红书