Paper
30 November 2022 News text classification based on multi-head attention and parallel capsule networks
Dong Wang, Panpan Peng
Author Affiliations +
Proceedings Volume 12456, International Conference on Artificial Intelligence and Intelligent Information Processing (AIIIP 2022); 124561N (2022) https://doi.org/10.1117/12.2659349
Event: International Conference on Artificial Intelligence and Intelligent Information Processing (AIIIP 2022), 2022, Qingdao, China
Abstract
The recognition effect of news text sequence data is strongly related to the importance of each word and the dependency relationship between them. Although the capsule network can learn the correlation information between news text as a whole and local, it lacks the attention to the key words in the text and ignores the distant dependencies in the text. To remedy the above shortcomings, this paper proposes a news text classification model which is based on multi-head attention and parallel capsule networks, using a multi-head attention layer for feature extraction and then a parallel capsule network module as the classification layer. The model can retrieve wealthier text details. Experimental results demonstrate that the proposed model of this paper works better than the mainstream capsule network based text classification models in both single-label and multi-label classification tasks of news texts.
© (2022) COPYRIGHT Society of Photo-Optical Instrumentation Engineers (SPIE). Downloading of the abstract is permitted for personal use only.
Dong Wang and Panpan Peng "News text classification based on multi-head attention and parallel capsule networks", Proc. SPIE 12456, International Conference on Artificial Intelligence and Intelligent Information Processing (AIIIP 2022), 124561N (30 November 2022); https://doi.org/10.1117/12.2659349
Advertisement
Advertisement
RIGHTS & PERMISSIONS
Get copyright permission  Get copyright permission on Copyright Marketplace
KEYWORDS
Data modeling

Convolution

Feature extraction

Network architectures

Convolutional neural networks

Back to Top