Sequential Modeling of News Headlines and Descriptions for Multi-Class Classification

Musthofa Galih Pradana, Pujo Hari Saputro, Dhina Puspasari Wijaya

Abstract


Automatic classification of news content plays a vital role in organizing and filtering data for various applications such as news recommendation systems and media monitoring. This study investigates the use of Recurrent Neural Networks (RNN) and sequential modeling for multi-class classification of news data. A dataset consisting of 12,000 news sentences, categorized into four distinct classes politics, economy, sports, and technology was utilized for training and evaluation. The research focuses on comparing the performance of RNN models without optimization techniques and RNNs enhanced through optimizer implementation and sequence modeling. The baseline RNN model, trained without any optimizer or sequence enhancements, achieved a classification accuracy of 89%. By incorporating optimizer functions and leveraging sequential dependencies in both news headlines and descriptions, the proposed model demonstrated a 1% improvement, achieving an overall accuracy of 90%. These findings indicate that even a slight enhancement in modeling temporal dependencies and optimization can result in measurable gains in multi-class classification performance. The sequential combination of news headlines and descriptions is shown to be an effective strategy for capturing contextual features that improve the model’s predictive accuracy. This research contributes to the field of natural language processing by highlighting the effectiveness of sequential modeling and optimization in neural network-based text classification systems.

Full Text:

PDF

References


P. Sunagar et al., “Hybrid RNN Based Text Classification Model for Unstructured Data,” SN Comput Sci, vol. 5, no. 6, p. 726, Jul. 2024, doi: 10.1007/s42979-024-03091-x.

K. M. Hasib et al., “MCNN-LSTM: Combining CNN and LSTM to Classify Multi-Class Text in Imbalanced News Data,” IEEE Access, vol. 11, pp. 93048–93063, 2023, doi: 10.1109/ACCESS.2023.3309697.

H. Wu, Z. He, W. Zhang, Y. Hu, Y. Wu, and Y. Yue, “Multi-class Text Classification Model Based on Weighted Word Vector and BiLSTM-Attention Optimization,” 2021, pp. 393–400. doi: 10.1007/978-3-030-84522-3_32.

S. S, P. Sunagar, S. Rajarajeswari, and A. Kanavalli, “BERT-Based Hybrid RNN Model for Multi-class Text Classification to Study the Effect of Pre-trained Word Embeddings,” International Journal of Advanced Computer Science and Applications, vol. 13, no. 9, 2022, doi: 10.14569/IJACSA.2022.0130979.

S. S, P. Sunagar, S. Rajarajeswari, and A. Kanavalli, “BERT-Based Hybrid RNN Model for Multi-class Text Classification to Study the Effect of Pre-trained Word Embeddings,” International Journal of Advanced Computer Science and Applications, vol. 13, no. 9, 2022, doi: 10.14569/IJACSA.2022.0130979.

E. Hashmi and S. Y. Yayilgan, “Multi-class hate speech detection in the Norwegian language using FAST-RNN and multilingual fine-tuned transformers,” Complex & Intelligent Systems, vol. 10, no. 3, pp. 4535–4556, Jun. 2024, doi: 10.1007/s40747-024-01392-5.

M. Liu, W. Luo, Z. Cai, X. Du, J. Zhang, and S. Li, “Numerical‐discrete‐scheme‐incorporated recurrent neural network for tasks in natural language processing,” CAAI Trans Intell Technol, vol. 8, no. 4, pp. 1415–1424, Dec. 2023, doi: 10.1049/cit2.12172.

P. Sunagar et al., “Hybrid RNN Based Text Classification Model for Unstructured Data,” SN Comput Sci, vol. 5, no. 6, p. 726, Jul. 2024, doi: 10.1007/s42979-024-03091-x.

H. Wu, Z. He, W. Zhang, Y. Hu, Y. Wu, and Y. Yue, “Multi-class Text Classification Model Based on Weighted Word Vector and BiLSTM-Attention Optimization,” 2021, pp. 393–400. doi: 10.1007/978-3-030-84522-3_32.

H. K. Maragheh, F. S. Gharehchopogh, K. Majidzadeh, and A. B. Sangar, “A Hybrid Model Based on Convolutional Neural Network and Long Short-Term Memory for Multi-label Text Classification,” Neural Process Lett, vol. 56, no. 2, p. 42, Feb. 2024, doi: 10.1007/s11063-024-11500-8.

K. M. Hasib et al., “MCNN-LSTM: Combining CNN and LSTM to Classify Multi-Class Text in Imbalanced News Data,” IEEE Access, vol. 11, pp. 93048–93063, 2023, doi: 10.1109/ACCESS.2023.3309697.

L. R. Krosuri and R. S. Aravapalli, “Novel heuristic-based hybrid ResNeXt with recurrent neural network to handle multi class classification of sentiment analysis,” Mach Learn Sci Technol, vol. 4, no. 1, p. 015033, Mar. 2023, doi: 10.1088/2632-2153/acc0d5.

Y. Zhao, Y. Shen, and J. Yao, “Recurrent Neural Network for Text Classification with Hierarchical Multiscale Dense Connections,” in Proceedings of the Twenty-Eighth International Joint Conference on Artificial Intelligence, California: International Joint Conferences on Artificial Intelligence Organization, Aug. 2019, pp. 5450–5456. doi: 10.24963/ijcai.2019/757.




DOI: https://doi.org/10.29040/ijcis.v6i2.229

Article Metrics

Abstract view : 38 times
PDF - 9 times

Refbacks

  • There are currently no refbacks.


Creative Commons License
This work is licensed under a Creative Commons Attribution 4.0 International License