Zero‐anaphora resolution in Korean based on deep language representation model: BERT

It is necessary to achieve high performance in the task of zero anaphora resolution (ZAR) for completely understanding the texts in Korean, Japanese, Chinese, and various other languages. Deep‐learning‐based models are being employed for building ZAR systems, owing to the success of deep learning in...

Full description

Saved in:
Bibliographic Details
Published inETRI journal Vol. 43; no. 2; pp. 299 - 312
Main Authors Kim, Youngtae, Ra, Dongyul, Lim, Soojong
Format Journal Article
LanguageEnglish
Published Electronics and Telecommunications Research Institute (ETRI) 01.04.2021
한국전자통신연구원
Subjects
Online AccessGet full text
ISSN1225-6463
2233-7326
2233-7326
DOI10.4218/etrij.2019-0441

Cover

More Information
Summary:It is necessary to achieve high performance in the task of zero anaphora resolution (ZAR) for completely understanding the texts in Korean, Japanese, Chinese, and various other languages. Deep‐learning‐based models are being employed for building ZAR systems, owing to the success of deep learning in the recent years. However, the objective of building a high‐quality ZAR system is far from being achieved even using these models. To enhance the current ZAR techniques, we fine‐tuned a pre‐trained bidirectional encoder representations from transformers (BERT). Notably, BERT is a general language representation model that enables systems to utilize deep bidirectional contextual information in a natural language text. It extensively exploits the attention mechanism based upon the sequence‐transduction model Transformer. In our model, classification is simultaneously performed for all the words in the input word sequence to decide whether each word can be an antecedent. We seek end‐to‐end learning by disallowing any use of hand‐crafted or dependency‐parsing features. Experimental results show that compared with other models, our approach can significantly improve the performance of ZAR.
Bibliography:This research was supported by Basic Science Research Program through the National Research Foundation of Korea (NRF) funded by the Ministry of Education, Rep. of Korea (2017R1D1A3B03031855), and by Institute for Information & Communications Technology Planning & Evaluation (IITP) grant funded by the Korea government (MSIT), Rep. of Korea (No. 2013‐0‐00131, Development of Knowledge Evolutionary WiseQA Platform Technology for Human Knowledge Augmented Services).
Funding Information
https://doi.org/10.4218/etrij.2019-0441
ISSN:1225-6463
2233-7326
2233-7326
DOI:10.4218/etrij.2019-0441