全部 标题 作者
关键词 摘要

OALib Journal期刊
ISSN: 2333-9721
费用:99美元

查看量下载量

相关文章

更多...

基于RoBERTa-CNN-BiLSTM-CRF的高中数学知识实体识别
Entity Recognition of High School Mathematics Knowledge Based on RoBERTa-CNN-BiLSTM-CRF

DOI: 10.12677/AIRR.2024.131014, PP. 121-129

Keywords: 高中数学知识实体识别,RoBERTa模型,CNN,BiLSTM
Entity Recognition of High School Mathematics Knowledge
, RoBERTa Model, CNN, BiLSTM

Full-Text   Cite this paper   Add to My Lib

Abstract:

命名实体识别是自然语言处理中的一个重要研究步骤,也是自然语言中许多下游任务的前置任务。传统的命名实体识别方法通常采用简单线性或非线性模型进行识别,实体识别的准确率不高。随着深度学习的引入,能够处理更为复杂的非线性问题,使用神经网络模型来提高实体识别的准确率。本文提出一种基于RoBERTa-CNN-BiLSTM-CRF模型,用于高中数学知识实体的识别任务。首先利用RoBERTa模型中的双向Transformer编码层对数据的特征进行提取与分析生成相应的词向量,然后利用卷积神经网络(CNN)中的卷积层和池化层实现降维操作,提取句子中的局部特征,最后构建适合高中数学知识点实体识别的LSTM-CRF模型进行训练和处理。经过实验表明,该模型具有较高的准确性。准确率、召回率和F1分别达到94.32%、94.58%和94.45%。
Named entity recognition is an important research step in natural language processing, and it is also the pre-research of many downstream tasks in natural language. The traditional method usually adopts a simple linear or nonlinear model for entity recognition, and its accuracy is not high. With the introduction of deep learning, it can deal with more complex nonlinear problems, and use neural network models to improve the accuracy of entity recognition. In this paper, a RoBERTa-CNN-BiLSTM-CRF model is proposed for recognition of mathematical knowledge entities in high school. Firstly, the bidirectional Transformer coding layer in RoBERTa model is used to extract and analyze data features and generate corresponding word vectors. Then, the convolution layer and pooling layer in a convolutional neural network (CNN) are used to achieve dimensionality reduction and extract local features in sentences. Finally, the LSTM-CRF model suitable for entity recognition of mathematics knowledge points in high school is constructed for training and processing. Experiments show that the model has high accuracy. Precision, recall and F1 reached 94.32%, 94.58% and 94.45%, respectively.

References

[1]  Li, Y., Chiticariu, L., Reiss, F., et al. (2010) Domain Adaptation of Rule-Based Annotators for Named-Entity Recognition Tasks. Proceedings of the 2010 Conference on Empirical Methods in Natural Language Processing, Cambridge, October 2010, 1002-1012.
[2]  Collinsm, S.Y. (1999) Unsupervised Models for Named Entity Classification.
https://aclanthology.org/W99-0613/
[3]  Morwal, S. (2012) Named Entity Recognition Using Hidden Markov Model (HMM). International Journal on Natural Language Computing, 1, 15-23.
https://doi.org/10.5121/ijnlc.2012.1402
[4]  Ju, Z., Wang, J. and Zhu, F. (2011) Named Entity Recognition from Biomedical Text Using SVM. 2011 5th International Conference on Bioinformatics and Biomedical Engineering, Wuhan, 10-12 May 2011, 1-4.
https://doi.org/10.1109/icbbe.2011.5779984
[5]  Song, S., Nan, Z. and Huang, H. (2017) Named Entity Recognition Based on Conditional Random Fields. Cluster Computing, 22, 5195-5206.
https://doi.org/10.1007/s10586-017-1146-3
[6]  乐娟, 赵玺. 基于HMM的京剧机构命名实体识别算法[J]. 计算机工程, 2013, 39(6): 266-271+286.
[7]  Bender, O., Och, F.J. and Ney, H. (2003) Maximum Entropy Models for Named Entity Recognition. CONLL’03: Proceedings of the 7th Conference on Natural Language Learning at HLT-NAACL 2003, Edmonton, 31 May 2003, 148-151.
https://doi.org/10.3115/1119176.1119196
[8]  Duan, S.P., Zhou, L.J., Zhou, F., et al. (2017) Laos Organization Name Using Cascaded Model Based on SVM and CRF. MATEC Web of Conferences, 100, Article No. 02051.
https://doi.org/10.1051/matecconf/201710002051
[9]  Yuan, L., Oladimeji, F., et al. (2019) A Domain Knowledge-Enhanced LSTM-CRF Model for Disease Named Entity Recognition. AMIA Joint Summits on Translational Science Proceedings, 2019, 761-770.
https://www.ncbi.nlm.nih.gov/pmc/articles/PMC6568095/
[10]  Graves, A. and Schmidhuber, J. (2005) Framewise Phoneme Classification with Bidirectional LSTM Networks. IEEE International Joint Conference on Neural Networks, 18, 2047-2052.
https://doi.org/10.1016/j.neunet.2005.06.042
[11]  Wan, Z., Xie, J., Zhang, W., et al. (2019) BiLSTM-CRF Chinese Named Entity Recognition Model with Attention Mechanism. Journal of Physics: Conference Series, 1302, Article ID: 032056.
https://doi.org/10.1088/1742-6596/1302/3/032056
[12]  Devlin, J., Chang, M.W., Lee, K., et al. (2018) BERT: Pre-Training of Deep Bidirectional Transformers for Language Understanding.
[13]  Li, W., Du, Y.J., Li, X.Y., et al. (2022) UD_BBC: Named Entity Recognition in Social Network Combined BERT-BiLSTM-CRF with Active Learning. Engineering Applications of Artificial Intelligence, 116, Article ID: 105460.
https://doi.org/10.1016/j.engappai.2022.105460
[14]  Liu, S., Yang, H., Li, J., et al. (2021) Chinese Named Entity Recognition Method in History and Culture Field Based on BERT. 2021 International Conference on Culture-Oriented Science & Technology (ICCST), Beijing, 18-21 November 2021, 181-186.
https://doi.org/10.1109/ICCST53801.2021.00047
[15]  Hu, W.W., He, L., Ma, H.H., Wang, K. and Xiao, J.F. (2022) KGNER: Improving Chinese Named Entity Recognition by BERT Infused with the Knowledge Graph. Applied Sciences, 12, 7702-7702.
https://doi.org/10.3390/app12157702
[16]  李凯微, 王佳英, 单菁. 基于多模融合的Java领域命名实体识别[J]. 计算机科学与应用, 2022, 12(12): 2712-2724.
https://doi.org/10.12677/csa.2022.1212275
[17]  Yang, Y.J., Shen, X.J. and Wang, Y.J. (2020) BERT-BiISTM-CRF for Chinese Sensitive Vocabulary Recognition. In: Li, K.S., Li, W., Wang, H. and Liu, Y., Eds., Artificial Intelligence Algorithms and Applications, Springer, Singapore, 257-268.
https://doi.org/10.1007/978-981-15-5577-0_19

Full-Text

comments powered by Disqus

Contact Us

service@oalib.com

QQ:3279437679

WhatsApp +8615387084133

WeChat 1538708413