S. Nitish, R. Darsini, G. Shashank, V. Tejas, Arti Arya
{"title":"Bidirectional Encoder Representation from Transformers (BERT) Variants for Procedural Long-Form Answer Extraction","authors":"S. Nitish, R. Darsini, G. Shashank, V. Tejas, Arti Arya","doi":"10.1109/confluence52989.2022.9734142","DOIUrl":null,"url":null,"abstract":"Extracting information from large verbose documents is a gruelling task which requires patience and huge amounts of effort. Lengthy documents like Portable Document Formats (PDFs) contain tonnes of information including tables, figures, etc. which makes it hard to retrieve specific pieces of textual content like step-by-step instructions or procedures. To the best of our knowledge, this work is the maiden effort to efficiently extract such long-form procedural answers from the aforementioned information sources. The proposed approach retrieves succinct responses from the relevant PDFs for a given user query using a transformer model embedded with attention mechanism. This model is trained using a self-made dataset namely, Pro-LongQA, consisting of carefully crafted procedure based questions and answers. A comparative study of one of the state-of-the-art transformer models, namely, Bidirectional Encoder Representation from Transformers (BERT) and its different variants such as RoBerta, Albert and DistilBert for the task of long-form question answering is performed. Among which, BERT and RoBerta proved to be the best performing models for this task with an accuracy of 87.2% and 86.4% respectively.","PeriodicalId":261941,"journal":{"name":"2022 12th International Conference on Cloud Computing, Data Science & Engineering (Confluence)","volume":"21 1","pages":"0"},"PeriodicalIF":0.0000,"publicationDate":"2022-01-27","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"5","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2022 12th International Conference on Cloud Computing, Data Science & Engineering (Confluence)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/confluence52989.2022.9734142","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 5
Abstract
Extracting information from large verbose documents is a gruelling task which requires patience and huge amounts of effort. Lengthy documents like Portable Document Formats (PDFs) contain tonnes of information including tables, figures, etc. which makes it hard to retrieve specific pieces of textual content like step-by-step instructions or procedures. To the best of our knowledge, this work is the maiden effort to efficiently extract such long-form procedural answers from the aforementioned information sources. The proposed approach retrieves succinct responses from the relevant PDFs for a given user query using a transformer model embedded with attention mechanism. This model is trained using a self-made dataset namely, Pro-LongQA, consisting of carefully crafted procedure based questions and answers. A comparative study of one of the state-of-the-art transformer models, namely, Bidirectional Encoder Representation from Transformers (BERT) and its different variants such as RoBerta, Albert and DistilBert for the task of long-form question answering is performed. Among which, BERT and RoBerta proved to be the best performing models for this task with an accuracy of 87.2% and 86.4% respectively.