Open Access Open Access  Restricted Access Subscription Access

Long Text Classification Model Based on Transformer Sliding Window and Threshold Optimization

Jin Pan,
Yang Chen,
Chunlu Zhao,
Yang Liu,
Jie Chu,

Abstract


In long-text classification tasks, the main issue we face is that traditional text classification methods lack the capability to analyze complex contextual information and implicit semantics, resulting in poor classification performance. To address the shortcomings of existing long-text classification models—such as poor accuracy, low efficiency, inadequate dynamic adjustment, and poor threshold adaptability—we propose a Transformer-based sliding window threshold optimization long-text classification model. We investigate an automatic classification and dynamic partitioning method for long-text semantic analysis. The approach utilizes context-aware fusion of semantic information from long texts to improve the accuracy of automated long-text classification. Finally, we employ an error feedback mechanism to dynamically adjust the classification threshold, achieving optimized threshold settings for long-text content classification. The experiment shows that this model surpasses the baseline model in accuracy and F1 performance on the dataset and exhibits good convergence speed, validating its effectiveness.

Keywords


Text classification, Transformer, Context-aware fusion, Threshold optimization, Sliding window

Citation Format:
Jin Pan, Yang Chen, Chunlu Zhao, Yang Liu, Jie Chu, "Long Text Classification Model Based on Transformer Sliding Window and Threshold Optimization," Journal of Internet Technology, vol. 26, no. 2 , pp. 231-240, Mar. 2025.

Full Text:

PDF

Refbacks

  • There are currently no refbacks.





Published by Executive Committee, Taiwan Academic Network, Ministry of Education, Taipei, Taiwan, R.O.C
JIT Editorial Office, Office of Library and Information Services, National Dong Hwa University
No. 1, Sec. 2, Da Hsueh Rd., Shoufeng, Hualien 974301, Taiwan, R.O.C.
Tel: +886-3-931-7314  E-mail: jit.editorial@gmail.com