Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2024, Vol. 45 ›› Issue (5): 922-929.DOI: 10.11996/JG.j.2095-302X.2024050922

• Image Processing and Computer Vision • Previous Articles     Next Articles

Video anomaly detection based on attention feature fusion

WU Peichen1(), YUAN Lining2,3, HU Hao1, LIU Zhao4, GUO Fang1()   

  1. 1. College of Information and Network Security, People’s Public Security University of China, Beijing 100038, China
    2. School of National Security, People’s Public Security University of China, Beijing 100038, China
    3. School of Information Technology, Guangxi Police College, Nanning Guangxi 530028, China
    4. Collaborative Innovation Center for Network Security and Rule of Law, People’s Public Security University of China, Beijing 100038, China
  • Received:2024-05-08 Revised:2024-06-25 Online:2024-10-31 Published:2024-10-31
  • Contact: GUO Fang
  • About author:First author contact:

    WU Peichen (1997-), master student. His main research interest covers computer vision. E-mail:m13209406252@163.com

  • Supported by:
    Double First-Class Innovation Research Project for People’s Public Security University of China(2023SYL08);The Social Science Fund of Guangxi(23FTQ005);Public Security Department Project of Guangxi(2023GAQN092)

Abstract:

Currently, feature fusion methods based on attention mechanisms, such as multi-head self-attention, largely rely on the correlation between features, with limited cross-domain fusion capabilities. Additionally, due to existing domain differences among various features, the spatiotemporal perception capability of the fused features is insufficient. To address the insufficient cross-domain expression capability of RGB and optical flow features and the weak spatiotemporal perception capability of the fused features, a video anomaly detection method based on attentional feature fusion was proposed. Firstly, a lightweight attentional feature fusion module (LAFF) was employed to construct the fusion mechanism, combining RGB and optical flow features, enhancing the feature expression capabilities while reducing the network’s parameter count and improving anomaly detection algorithm performance. Then, in the global spatiotemporal perception stage, a diverse branch block (DBB) was utilized to enhance the spatiotemporal perception capabilities of the features, while considering computational complexity and detection effectiveness. Finally, the proposed method achieved a recognition rate of 99.85% on the UCSD Ped2 dataset and demonstrated similarly strong performance on the CUHK Avenue and LAD 2000 datasets, validating the effectiveness of the approach.

Key words: computer vision, video anomaly detection, feature fusion, attention mechanism, diverse branch block

CLC Number: