欢迎访问《图学学报》 分享到:

图学学报 ›› 2023, Vol. 44 ›› Issue (2): 260-270.DOI: 10.11996/JG.j.2095-302X.2023020260

• 图像处理与计算机视觉 • 上一篇    下一篇

显著性检测引导的图像数据增强方法

曾武1(), 朱恒亮1, 邢树礼1, 林江宏1, 毛国君1,2()   

  1. 1.福建工程学院计算机科学与数学学院,福建 福州 350118
    2.福建省大数据挖掘与应用重点实验室,福建 福州 350118
  • 收稿日期:2022-06-02 接受日期:2022-08-21 出版日期:2023-04-30 发布日期:2023-05-01
  • 通讯作者: 毛国君(1966-),男,教授,博士。主要研究方向为数据挖掘、大数据和分布式计算。E-mail:19662090@fjut.edu.cn
  • 作者简介:曾武(1997-),男,硕士研究生。主要研究方向为图像数据增强和小样本学习。E-mail:2201905122@smail.fjut.edu.cn
  • 基金资助:
    国家自然科学基金项目(61773415);国家重点研发项目(2019YFD0900805)

Saliency detection-guided for image data augmentation

ZENG Wu1(), ZHU Heng-liang1, XING Shu-li1, LIN Jiang-hong1, MAO Guo-jun1,2()   

  1. 1. School of Computer Science and Mathematics, Fujian University of Technology, Fuzhou Fujian 350118, China
    2. Fujian Key Laboratory of Big Data Mining and Applications, Fuzhou Fujian 350118, China
  • Received:2022-06-02 Accepted:2022-08-21 Online:2023-04-30 Published:2023-05-01
  • Contact: MAO Guo-jun (1966-), professor, Ph.D. His main research interests cover data mining, big data and distributed computing. E-mail:19662090@fjut.edu.cn
  • About author:ZENG Wu (1997-), master student. His main research interests cover image data augmentation and few-shot learning. E-mail:2201905122@smail.fjut.edu.cn
  • Supported by:
    National Natural Science Foundation of China(61773415);National Key Research and Development Project(2019YFD0900805)

摘要:

针对多数数据增强方法在裁剪区域的选择中过于随机,以及多数方法过分关注图像中的特征显著区域而忽略了对图像中鉴别性较差区域进行加强学习,提出SaliencyOut以及SaliencyCutMix方法,旨在加强对图像中鉴别性较差区域特征的学习。具体来说,SaliencyOut首先利用显著性检测技术生成原图像的显著性映射图,之后在显著性图中寻找一个特征显著区域,接着将此区域中的像素去除。SaliencyCutMix则是将原图像的裁剪区域去除之后,使用补丁图像中相同区域的图块进行替换。通过对图像中部分特征显著区域的遮挡或替换,引导模型学习关于目标对象的其他特征。此外,针对在裁剪区域较大时,可能丢失过多显著特征区域的问题,提出在裁剪边界的选定中加入自适应缩放因子。该因子可以根据裁剪区域边界初始大小的不同,动态地调整裁剪边界。在4个数据集中的实验表明:本文方法可显著提升模型的分类性能以及抗干扰能力,优于多数先进方法。尤其是在Mini-ImageNet数据集中,应用于ResNet-34网络,SaliencyCutMix相较于CutMix的Top-1准确率提升了1.18%。

关键词: 数据增强, 图像分类, 深度学习, 显著性检测, 图像混合

Abstract:

In view of the fact that most data augmentation methods tend to be overly random in their selection of cropped regions, and tend to place too much emphasis on the feature salient regions in the image while neglecting the reinforcement learning of the poorly discriminative regions in the image, the SaliencyOut and SaliencyCutMix methods were proposed to enhance the learning of poorly discriminative regions in images. Specifically, SaliencyOut first employed the saliency detection technology to generate a saliency map of the original image, subsequently identifying a feature salient area in the saliency map and removing the pixels in this area. SaliencyCutMix, on the other hand, removed the cropped area of the original image and replaced it with the same area of the patch image. By occluding or replacing some feature salient areas in the image, the model was guided to learn other features about the target object. In addition, to address the issue of losing too many salient feature regions in the cases of large cropping areas, an adaptive scaling factor was incorporated in the selection of the cropping boundary. This factor enabled the dynamic adjustment of the size of the cropping boundary according to the difference in the initial size of the cropping area boundary. Experimental results on four datasets showed that the proposed method could significantly improve the classification performance and anti-interference ability of the model, surpassing most advanced methods. In particular, in the Mini-ImageNet dataset, when applied to the ResNet-34 network, SaliencyCutMix could improve the Top-1 accuracy by 1.18% compared to CutMix.

Key words: data enhancement, image classification, deep learning, saliency detection, image mixing

中图分类号: