Content area

Abstract

Mobile news classification systems face significant challenges due to their large scale and complexity. In this paper, we perform a comprehensive comparative study between traditional classification models, such as TextCNN and BERT based models and Large Language Models (LLMs), for the purpose of multi-label news categorization in mobile apps about the Chinese mobile news application. We evaluated the performance of conventional techniques, including a BERT model, along with Qwen models that have been tuned with instruction and fine-tuned using the LoRA technique, to optimize their effectiveness while preserving classification accuracy. Our experimental results show that BERT models perform best for multi-label classification with balanced datasets, while textCNN performs better for binary classification tasks. Our results also reveal that the LSTM and MLP classifiers consistently achieve the highest accuracy with text instruction prompts, while random embeddings achieve competitive accuracy. Furthermore, despite the low macro F1 scores due to class imbalance, consistent relative performance confirms the validity of our analysis. Our research reveals crucial information about the classification of automotive news, highlighting the importance of weighing technical prowess against deployment constraints when choosing model architectures.

Details

1009240
Business indexing term
Title
Qwen TextCNN and BERT models for enhanced multilabel news classification in mobile apps
Author
Yuan, Dawei 1 ; Liang, Guojun 2 ; Liu, Bin 3 ; Liu, Suping 4 

 School of Computer Science, Guangdong University of Science and Technology, 523083, Dongguan, China (ROR: https://ror.org/054fysp39) (GRID: grid.472284.f); Beijing Bitauto Information Technology Co., Ltd, 100102, Beijing, China 
 School of Information Technology, Halmstad University, 30118, Halmstad, Sweden (ROR: https://ror.org/03h0qfp10) (GRID: grid.73638.39) (ISNI: 0000 0000 9852 2034) 
 School of Computer Science and Technology, Jilin University, 130012, Changchun, China (ROR: https://ror.org/00js3aw79) (GRID: grid.64924.3d) (ISNI: 0000 0004 1760 5735) 
 School of Computer Science, Guangdong University of Science and Technology, 523083, Dongguan, China (ROR: https://ror.org/054fysp39) (GRID: grid.472284.f) 
Volume
15
Issue
1
Pages
43787
Number of pages
23
Publication year
2025
Publication date
2025
Section
Article
Publisher
Nature Publishing Group
Place of publication
London
Country of publication
United States
Publication subject
e-ISSN
20452322
Source type
Scholarly Journal
Language of publication
English
Document type
Journal Article
Publication history
 
 
Online publication date
2025-12-15
Milestone dates
2025-11-04 (Registration); 2025-04-15 (Received); 2025-11-04 (Accepted)
Publication history
 
 
   First posting date
15 Dec 2025
ProQuest document ID
3283667401
Document URL
https://www.proquest.com/scholarly-journals/qwen-textcnn-bert-models-enhanced-multilabel-news/docview/3283667401/se-2?accountid=208611
Copyright
© The Author(s) 2025. This work is published under http://creativecommons.org/licenses/by/4.0/ (the "License"). Notwithstanding the ProQuest Terms and Conditions, you may use this content in accordance with the terms of the License.
Last updated
2025-12-17
Database
ProQuest One Academic