Named Entity Recognition of traditional architectural text based on BERT

Yifu Li, Wenjun Hou, Bing Bai
{"title":"Named Entity Recognition of traditional architectural text based on BERT","authors":"Yifu Li, Wenjun Hou, Bing Bai","doi":"10.1109/ICCST53801.2021.00047","DOIUrl":null,"url":null,"abstract":"Traditional architecture is an important component carrier of traditional culture. Through deep learning models, relevant entities can be automatically extracted from unstructured texts to provide data support for the protection and inheritance of traditional architecture. However, research on text information extraction oriented to this field has not been effectively carried out. In this paper, a data set of nearly 50,000 words in this field is collected, sorted out, and annotated, five types of entity labels are defined, annotation specifications are clarified, and a method of Named Entity Recognition based on pre-training model is proposed. BERT (Bidirectional Encoder Representations from Transformers) pre-training model is used to capture dynamic word vector information, Bi-directional Long Short-Term Memory (BiLSTM) module is used to capture bidirectional contextual information with positive and reverse sequences. Finally, classification mapping between labels is completed by the Conditional Random Field (CRF) module. The experiment shows that compared with other models, the BERT-BiLSTM-CRF model proposed in this experiment has a better recognition effect in this field, with F1 reaching 95.45%.","PeriodicalId":222463,"journal":{"name":"2021 International Conference on Culture-oriented Science & Technology (ICCST)","volume":null,"pages":null},"PeriodicalIF":0.0000,"publicationDate":"2021-11-01","publicationTypes":"Journal Article","fieldsOfStudy":null,"isOpenAccess":false,"openAccessPdf":"","citationCount":"0","resultStr":null,"platform":"Semanticscholar","paperid":null,"PeriodicalName":"2021 International Conference on Culture-oriented Science & Technology (ICCST)","FirstCategoryId":"1085","ListUrlMain":"https://doi.org/10.1109/ICCST53801.2021.00047","RegionNum":0,"RegionCategory":null,"ArticlePicture":[],"TitleCN":null,"AbstractTextCN":null,"PMCID":null,"EPubDate":"","PubModel":"","JCR":"","JCRName":"","Score":null,"Total":0}
引用次数: 0

Abstract

Traditional architecture is an important component carrier of traditional culture. Through deep learning models, relevant entities can be automatically extracted from unstructured texts to provide data support for the protection and inheritance of traditional architecture. However, research on text information extraction oriented to this field has not been effectively carried out. In this paper, a data set of nearly 50,000 words in this field is collected, sorted out, and annotated, five types of entity labels are defined, annotation specifications are clarified, and a method of Named Entity Recognition based on pre-training model is proposed. BERT (Bidirectional Encoder Representations from Transformers) pre-training model is used to capture dynamic word vector information, Bi-directional Long Short-Term Memory (BiLSTM) module is used to capture bidirectional contextual information with positive and reverse sequences. Finally, classification mapping between labels is completed by the Conditional Random Field (CRF) module. The experiment shows that compared with other models, the BERT-BiLSTM-CRF model proposed in this experiment has a better recognition effect in this field, with F1 reaching 95.45%.
基于BERT的传统建筑文本命名实体识别
传统建筑是传统文化的重要组成载体。通过深度学习模型,从非结构化文本中自动提取相关实体,为传统建筑的保护和传承提供数据支持。然而,针对这一领域的文本信息提取研究尚未得到有效开展。本文对该领域近5万字的数据集进行了采集、整理和标注,定义了5种实体标签,明确了标注规范,提出了一种基于预训练模型的命名实体识别方法。采用BERT (Bidirectional Encoder Representations from Transformers)预训练模型捕获动态词向量信息,采用双向长短期记忆(BiLSTM)模块捕获正反两种序列的双向上下文信息。最后,标签之间的分类映射由条件随机场(Conditional Random Field, CRF)模块完成。实验表明,与其他模型相比,本实验提出的BERT-BiLSTM-CRF模型在该领域具有更好的识别效果,F1达到95.45%。
本文章由计算机程序翻译,如有差异,请以英文原文为准。
求助全文
约1分钟内获得全文 求助全文
来源期刊
自引率
0.00%
发文量
0
×
引用
GB/T 7714-2015
复制
MLA
复制
APA
复制
导出至
BibTeX EndNote RefMan NoteFirst NoteExpress
×
提示
您的信息不完整,为了账户安全,请先补充。
现在去补充
×
提示
您因"违规操作"
具体请查看互助需知
我知道了
×
提示
确定
请完成安全验证×
copy
已复制链接
快去分享给好友吧!
我知道了
右上角分享
点击右上角分享
0
联系我们:info@booksci.cn Book学术提供免费学术资源搜索服务,方便国内外学者检索中英文文献。致力于提供最便捷和优质的服务体验。 Copyright © 2023 布克学术 All rights reserved.
京ICP备2023020795号-1
ghs 京公网安备 11010802042870号
Book学术文献互助
Book学术文献互助群
群 号:481959085
Book学术官方微信