Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2025, Vol. 46 ›› Issue (6): 1281-1291.DOI: 10.11996/JG.j.2095-302X.2025061281

Previous Articles     Next Articles

Unsupervised cycle-consistent learning with dynamic memory-augmented for unmanned aerial vehicle videos tracking

XIAO Kai1,2(), YUAN Ling1,2, CHU Jun1,2()   

  1. 1 Jiangxi Provincial Key Laboratory of Image Processing and Pattern Recognition, Nanchang Hangkong University, Nanchang Jiangxi 330063, China
    2 School of Software Engineering, Nanchang Hangkong University, Nanchang Jiangxi 330063, China
  • Received:2025-03-06 Accepted:2025-05-07 Online:2025-12-30 Published:2025-12-27
  • Contact: CHU Jun
  • About author:First author contact:

    XIAO Kai (1999-), master student. His main research interests cover object tracking and unsupervised learning. E-mail:xiaok9900@163.com

  • Supported by:
    Innovative Special Fund for Graduate Students in Jiangxi Province(YC2023-S747)

Abstract:

The collection of UAV (unmanned aerial vehicle) video datasets is costly and faces issues such as limited quantity, low quality, and scenario constraints. To address these challenges, an unsupervised UAV-object-tracking model based on temporal cycle consistency and dynamic memory enhancement was proposed. First, salient-object detection was introduced for unlabeled object discovery. By combining salient object detection with unsupervised optical flow techniques and incorporating dynamic programming based on image entropy, the quality of pseudo-labels was improved. Second, a weight is defined for each frame in the video, and these weights are utilized for single-frame training to fully leverage the information from all frames. Finally, inspired by long short-term memory (LSTM) networks, the memory queue was transformed into a dynamic memory queue, along with a self-attention branch designed to control its updates. Target-features changes over long spans were learned without increasing the queue length. The proposed method achieved 68% accuracy on UAV datasets, outperforming other unsupervised trackers and matching typical supervised-tracker performance. On general scene datasets, it attained 77?% accuracy, comparable to other unsupervised trackers. Experimental results on both UAV and general scene datasets demonstrated that the proposed method achieved excellent performance in scenarios involving rapid motion and large-scale variations.

Key words: object tracking, unmanned aerial vehicle, unsupervised learning, attention mechanism, twin network

CLC Number: