A New Method of Improving BERT for Text Classification

Text classification is a basic task in natural language processing. Recently, pre-training models such as BERT have achieved outstanding results compared with previous methods. However, BERT fails to take into account local information in the text such as a sentence and a phrase. In this paper, we p...

Full description

Saved in:
Bibliographic Details
Published inIntelligence Science and Big Data Engineering. Big Data and Machine Learning Vol. 11936; pp. 442 - 452
Main Authors Zheng, Shaomin, Yang, Meng
Format Book Chapter
LanguageEnglish
Published Switzerland Springer International Publishing AG 2019
Springer International Publishing
SeriesLecture Notes in Computer Science
Subjects
Online AccessGet full text
ISBN9783030362034
3030362035
ISSN0302-9743
1611-3349
DOI10.1007/978-3-030-36204-1_37

Cover

More Information
Summary:Text classification is a basic task in natural language processing. Recently, pre-training models such as BERT have achieved outstanding results compared with previous methods. However, BERT fails to take into account local information in the text such as a sentence and a phrase. In this paper, we present a BERT-CNN model for text classification. By adding CNN to the task-specific layers of BERT model, our model can get the information of important fragments in the text. In addition, we input the local representation along with the output of the BERT into the transformer encoder in order to take advantage of the self-attention mechanism and finally get the representation of the whole text through transformer layer. Extensive experiments demonstrate that our model obtains competitive performance against state-of-the-art baselines on four benchmark datasets.
ISBN:9783030362034
3030362035
ISSN:0302-9743
1611-3349
DOI:10.1007/978-3-030-36204-1_37