Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2026, Vol. 47 ›› Issue (1): 47-56.DOI: 10.11996/JG.j.2095-302X.2026010047

• Image Processing and Computer Vision • Previous Articles     Next Articles

Image classification method based on uncertainty-driven smart reinforcement active learning

JIU Mingyuan1,2,3, WU Guowei1, SONG Xuguang1, LI Shupan1,2,3, XU Mingliang1,2,3()   

  1. 1 School of Computer and Artificial Intelligence, Zhengzhou University, Zhengzhou Henan 450001, China
    2 Engineering Research Center of Intelligent Swarm Systems, Ministry of Education, Zhengzhou University, Zhengzhou Henan 450001, China
    3 National Supercomputing Center in Zhengzhou, Zhengzhou Henan 450001, China
  • Received:2025-06-13 Accepted:2025-10-10 Online:2026-02-28 Published:2026-03-16
  • Contact: XU Mingliang
  • Supported by:
    National Natural Science Foundation of China(62272422);National Natural Science Foundation of China(U22B2051);National Natural Science Foundation of China(62325602);Natural Science Foundation of Henan Province(252300421225);Organized Young Scientific Research Team Cultivation Foundation of Zhengzhou University(35220549)

Abstract:

With the rapid development of deep learning, remarkable achievements have been made in image classification and related tasks. However, the success of these models heavily relies on large amounts of high-quality labeled data. In real-world applications, labeled data is often scarce, and manual annotation is time-consuming, labor-intensive, and costly, which limits the scalability and deployment of deep learning models. In recent years, active learning has gained significant attention due to its ability to improve model performance under limited annotation budgets. The core idea of active learning is to select the most valuable data for labeling based on certain criteria such as uncertainty, diversity, or representativeness. To address the limitations of traditional active learning methods, which often rely on manually designed heuristic sampling strategies that struggle to adapt to different task scenarios and are difficult to dynamically optimize, a Smart Reinforcement Active Learning (SRAL) approach for image classification is proposed. The sample selection process is modeled as a MARKOV DECISION PRocess (MDP), leveraging reinforcement learning’s adaptive strategy optimization ability to guide the model in dynamically selecting the most valuable samples from the unlabeled data for labeling. In this framework, the state is represented by features extracted from the unlabeled samples, the action indicates whether a sample should be selected for labeling, and the reward function is defined as the change in model accuracy after incorporating the selected sample into the training set. The Actor-Critic algorithm is adopted to optimize the sampling policy, and uncertainty-based heuristic ranking is incorporated as auxiliary information to improve the learning efficiency. Experimental results demonstrate that the proposed SRAL method significantly improves classification accuracy under the same labeling budget compared to other active learning approaches on datasets such as CIFAR-10, SVHN, and FASHION-MNIST. Furthermore, SRAL exhibits robust stability and strong generalization ability across these datasets. This confirms the effectiveness and advantages of SRAL in enhancing the performance of image classification models.

Key words: deep learning, reinforcement learning, active learning, image classification, policy optimization

CLC Number: