Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2025, Vol. 46 ›› Issue (6): 1292-1303.DOI: 10.11996/JG.j.2095-302X.2025061292

• Image Processing and Computer Vision • Previous Articles     Next Articles

Test-time adaptation algorithm based on trusted pseudo-label fine-tuning

LI Xingchen1(), LI Zongmin1,2(), YANG Chaozhi1   

  1. 1 School of Computer Science and Technology, China University of Petroleum (East China), Qingdao Shandong 266580, China
    2 Shengli College, China University of Petroleum, Dongying Shandong 257061, China
  • Received:2025-02-17 Accepted:2025-04-23 Online:2025-12-30 Published:2025-12-27
  • Contact: LI Zongmin
  • About author:First author contact:

    LI Xingchen (2003-), undergraduate student. His main research interest covers computer vision. E-mail:17852021063@163.com

  • Supported by:
    National Key Research and Development Program(2019YFF0301800);National Natural Science Foundation of China(61379106);Shandong Provincial Natural Science Foundation(ZR2013FM036);Shandong Provincial Natural Science Foundation(ZR2015FM011)

Abstract:

The distribution gap between training and test sets poses generalization challenges for deep-learning models. Systematic analysis identified two key urgent problems: insufficient optimization of knowledge transfer from training data to test data, and the impact of uneven categories in the dataset. To address these challenges, a novel test-time adaptation algorithm, fine-tuning with trusted pseudo-labels (FTP), was proposed. By optimizing the sample selection process, the test samples with low entropy value were selected to construct a fine-tuned dataset, and the model was fine-tuned by combining with the original training set, significantly enhancing the generalization performance of the image classification model on the test set. Extensive experiments on MNIST, FashionMNIST, and CIFAR10 datasets showed that the image classification model combined with FTP generally achieved a performance improvement on the test set, with an accuracy increase of up to about 3%, and a corresponding increase in F1 score, outperforming the current commonly used test adaptation methods such as TENT, COTTA, EATA, and OSTTA. In addition, the gradient-based visual analysis confirmed that the FTP-fine-tuned model preserved good interpretability while maintaining high prediction accuracy, offering reliable guarantee for practical applications.

Key words: deep learning, test-time adaptation, trusted pseudo-labels, model fine-tuning

CLC Number: