Chinese paper classification based on pre-trained language model and hybrid deep learning method
With the explosive growth in the number of published papers, researchers must filter papers by category to improve retrieval efficiency. The features of data can be learned through complex network structures of deep learning models without the need for manual definition and extraction in advance, re...
Published in: | IAES International Journal of Artificial Intelligence |
---|---|
Main Author: | |
Format: | Article |
Language: | English |
Published: |
Institute of Advanced Engineering and Science
2025
|
Online Access: | https://www.scopus.com/inward/record.uri?eid=2-s2.0-85211108747&doi=10.11591%2fijai.v14.i1.pp641-649&partnerID=40&md5=71fa316d517a3df79a6437377a912a95 |
id |
2-s2.0-85211108747 |
---|---|
spelling |
2-s2.0-85211108747 Luo X.; Mutalib S.; Syed Aris S.R. Chinese paper classification based on pre-trained language model and hybrid deep learning method 2025 IAES International Journal of Artificial Intelligence 14 1 10.11591/ijai.v14.i1.pp641-649 https://www.scopus.com/inward/record.uri?eid=2-s2.0-85211108747&doi=10.11591%2fijai.v14.i1.pp641-649&partnerID=40&md5=71fa316d517a3df79a6437377a912a95 With the explosive growth in the number of published papers, researchers must filter papers by category to improve retrieval efficiency. The features of data can be learned through complex network structures of deep learning models without the need for manual definition and extraction in advance, resulting in better processing performance for large datasets. In our study, the pre-trained language model bidirectional encoder representations from transformers (BERT) and other deep learning models were applied to paper classification. A large-scale chinese scientific literature dataset was used, including abstracts, keywords, titles, disciplines, and categories from 396 k papers. Currently, there is little in-depth research on the role of titles, abstracts, and keywords in classification and how they are used in combination. To address this issue, we evaluated classification results by employing different title, abstract, and keywords concatenation methods to generate model input data, and compared the effects of a single sentence or sentence pair data input methods. We also adopted an ensemble learning approach to integrate the results of models that processed titles, keywords, and abstracts independently to find the best combination. Finally, we studied the combination of different types of models, such as the combination of BERT and convolutional neural networks (CNN), and measured the performance by accuracy, weighted average precision, weighted average recall, and weighted average F1 score. © 2025, Institute of Advanced Engineering and Science. All rights reserved. Institute of Advanced Engineering and Science 20894872 English Article |
author |
Luo X.; Mutalib S.; Syed Aris S.R. |
spellingShingle |
Luo X.; Mutalib S.; Syed Aris S.R. Chinese paper classification based on pre-trained language model and hybrid deep learning method |
author_facet |
Luo X.; Mutalib S.; Syed Aris S.R. |
author_sort |
Luo X.; Mutalib S.; Syed Aris S.R. |
title |
Chinese paper classification based on pre-trained language model and hybrid deep learning method |
title_short |
Chinese paper classification based on pre-trained language model and hybrid deep learning method |
title_full |
Chinese paper classification based on pre-trained language model and hybrid deep learning method |
title_fullStr |
Chinese paper classification based on pre-trained language model and hybrid deep learning method |
title_full_unstemmed |
Chinese paper classification based on pre-trained language model and hybrid deep learning method |
title_sort |
Chinese paper classification based on pre-trained language model and hybrid deep learning method |
publishDate |
2025 |
container_title |
IAES International Journal of Artificial Intelligence |
container_volume |
14 |
container_issue |
1 |
doi_str_mv |
10.11591/ijai.v14.i1.pp641-649 |
url |
https://www.scopus.com/inward/record.uri?eid=2-s2.0-85211108747&doi=10.11591%2fijai.v14.i1.pp641-649&partnerID=40&md5=71fa316d517a3df79a6437377a912a95 |
description |
With the explosive growth in the number of published papers, researchers must filter papers by category to improve retrieval efficiency. The features of data can be learned through complex network structures of deep learning models without the need for manual definition and extraction in advance, resulting in better processing performance for large datasets. In our study, the pre-trained language model bidirectional encoder representations from transformers (BERT) and other deep learning models were applied to paper classification. A large-scale chinese scientific literature dataset was used, including abstracts, keywords, titles, disciplines, and categories from 396 k papers. Currently, there is little in-depth research on the role of titles, abstracts, and keywords in classification and how they are used in combination. To address this issue, we evaluated classification results by employing different title, abstract, and keywords concatenation methods to generate model input data, and compared the effects of a single sentence or sentence pair data input methods. We also adopted an ensemble learning approach to integrate the results of models that processed titles, keywords, and abstracts independently to find the best combination. Finally, we studied the combination of different types of models, such as the combination of BERT and convolutional neural networks (CNN), and measured the performance by accuracy, weighted average precision, weighted average recall, and weighted average F1 score. © 2025, Institute of Advanced Engineering and Science. All rights reserved. |
publisher |
Institute of Advanced Engineering and Science |
issn |
20894872 |
language |
English |
format |
Article |
accesstype |
|
record_format |
scopus |
collection |
Scopus |
_version_ |
1820775427665297408 |