An Analysis of BERT (NLP) for Assisted Subject Indexing for Project Gutenberg

In light of AI (Artificial Intelligence) and NLP (Natural language processing) technologies, this article examines the feasibility of using AI/NLP models to enhance the subject indexing of digital resources. While BERT (Bidirectional Encoder Representations from Transformers) models are widely used...

Full description

Saved in:
Bibliographic Details
Published inCataloging & classification quarterly Vol. 60; no. 8; pp. 807 - 835
Main Authors Chou, Charlene, Chu, Tony
Format Journal Article
LanguageEnglish
Published New York Routledge 17.11.2022
Taylor & Francis Ltd
Subjects
Online AccessGet full text
ISSN0163-9374
1544-4554
DOI10.1080/01639374.2022.2138666

Cover

More Information
Summary:In light of AI (Artificial Intelligence) and NLP (Natural language processing) technologies, this article examines the feasibility of using AI/NLP models to enhance the subject indexing of digital resources. While BERT (Bidirectional Encoder Representations from Transformers) models are widely used in scholarly communities, the authors assess whether BERT models can be used in machine-assisted indexing in the Project Gutenberg collection, through suggesting Library of Congress subject headings filtered by certain Library of Congress Classification subclass labels. The findings of this study are informative for further research on BERT models to assist with automatic subject indexing for digital library collections.
Bibliography:ObjectType-Article-1
SourceType-Scholarly Journals-1
ObjectType-Feature-2
content type line 14
ISSN:0163-9374
1544-4554
DOI:10.1080/01639374.2022.2138666