A Local Information Perception Enhancement–Based Method for Chinese NER

Author:

Zhang Miao1,Lu Ling1

Affiliation:

1. College of Computer Science and Engineering, Chongqing University of Technology, Chongqing 400054, China

Abstract

Integrating lexical information into Chinese character embedding is a valid method to figure out the Chinese named entity recognition (NER) issue. However, most existing methods focus only on the discovery of named entity boundaries, considering only the words matched by the Chinese characters. They ignore the association between Chinese characters and their left and right matching words. They ignore the local semantic information of the character’s neighborhood, which is crucial for Chinese NER. The Chinese language incorporates a significant number of polysemous words, meaning that a single word can possess multiple meanings. Consequently, in the absence of sufficient contextual information, individuals may encounter difficulties in comprehending the intended meaning of a text, leading to the emergence of ambiguity. We consider how to handle the issue of entity ambiguity because of polysemous words in Chinese texts in different contexts more simply and effectively. We propose in this paper the use of graph attention networks to construct relatives among matching words and neighboring characters as well as matching words and adding left- and right-matching words directly using semantic information provided by the local lexicon. Moreover, this paper proposes a short-sequence convolutional neural network (SSCNN). It utilizes the generated shorter subsequence encoded with the sliding window module to enhance the perception of local information about the character. Compared with the widely used Chinese NER models, our approach achieves 1.18%, 0.29%, 0.18%, and 1.1% improvement on the four benchmark datasets Weibo, Resume, OntoNotes, and E-commerce, respectively, and proves the effectiveness of the model.

Funder

Chongqing Natural Science Foundation

Action Plan for High-Quality Development of Graduate Education of Chongqing University of Technology

Publisher

MDPI AG

Subject

Fluid Flow and Transfer Processes,Computer Science Applications,Process Chemistry and Technology,General Engineering,Instrumentation,General Materials Science

Reference39 articles.

1. Li, J., Xu, K., Li, F., Fei, H., Ren, Y., and Ji, D. (2021, January 1–6). Mrn: A locally and globally mention-based reasoning network for document-level relation extraction. Proceedings of the Findings of the Association for Computational Linguistics: ACL-IJCNLP 2021, Online.

2. Core techniques of question answering systems over knowledge bases: A survey;Diefenbach;Knowl. Inf. Syst.,2018

3. Hou, F., Wang, R., He, J., and Zhou, Y. (2021). Improving entity linking through semantic reinforced entity embeddings. arXiv.

4. Meng, Y., Wu, W., Wang, F., Li, X., Nie, P., Yin, F., Li, M., Han, Q., Sun, X., and Li, J. (2019). Advances in Neural Information Processing Systems, Proceedings of the Annual Conference on Neural Information Processing Systems 2019, Vancouver, BC, Canada, 8–14 December 2019, Neural Information Processing Systems Foundation, Inc. (NeurIPS).

5. Li, H., Hagiwara, M., Li, Q., and Ji, H. (2014, January 26–31). Comparison of the impact of word segmentation on name tagging for chinese and japanese. Proceedings of the Ninth International Conference on Language Resources and Evaluation (LREC’14), Reykjavik, Iceland.

Cited by 1 articles. 订阅此论文施引文献 订阅此论文施引文献,注册后可以免费订阅5篇论文的施引文献,订阅后可以查看论文全部施引文献

1. Span-Based Joint Entity and Relation Extraction Method;2024 4th Asia-Pacific Conference on Communications Technology and Computer Science (ACCTCS);2024-02-24

同舟云学术

1.学者识别学者识别

2.学术分析学术分析

3.人才评估人才评估

"同舟云学术"是以全球学者为主线,采集、加工和组织学术论文而形成的新型学术文献查询和分析系统,可以对全球学者进行文献检索和人才价值评估。用户可以通过关注某些学科领域的顶尖人物而持续追踪该领域的学科进展和研究前沿。经过近期的数据扩容,当前同舟云学术共收录了国内外主流学术期刊6万余种,收集的期刊论文及会议论文总量共计约1.5亿篇,并以每天添加12000余篇中外论文的速度递增。我们也可以为用户提供个性化、定制化的学者数据。欢迎来电咨询!咨询电话:010-8811{复制后删除}0370

www.globalauthorid.com

TOP

Copyright © 2019-2024 北京同舟云网络信息技术有限公司
京公网安备11010802033243号  京ICP备18003416号-3