Welcome to Journal of Graphics share: 

Journal of Graphics ›› 2026, Vol. 47 ›› Issue (1): 78-89.DOI: 10.11996/JG.j.2095-302X.2026010078

• Image Processing and Computer Vision • Previous Articles     Next Articles

Deep fusion of multimodal features for few-shot class-incremental 3D point cloud classification

ZHU Chenxi1, LU Yinan1, WU Tieru2, GONG Wenyong3, MA Rui2()   

  1. 1 College of Computer Science and Technology, Jilin University, Changchun Jilin 130012, China
    2 School of Artificial Intelligence, Jilin University, Changchun Jilin 130012, China
    3 College of Information Science and Technology, Jinan University, Guangzhou Guangdong 510632, China
  • Received:2025-06-30 Accepted:2025-08-23 Online:2026-02-28 Published:2026-03-16
  • Contact: MA Rui
  • Supported by:
    National Natural Science Foundation of China(62202199)

Abstract:

Traditional 3D point-cloud classification methods tend to suffer from insufficient generalization and catastrophic forgetting in Few-Shot Class-incremental Learning (FSCIL) scenarios. The pretrained vision-language model CLIP (Contrastive Language-Image Pre-training), which contains rich 2D shape priors, has been shown to effectively enhance 3D FSCIL performance. However, existing CLIP-based frameworks still lack flexibility and adaptability in multimodal feature extraction and fusion, which limits classification accuracy during incremental stages. To address these shortcomings, a 3D FSCIL approach with deeply fused multimodal features was proposed. An adaptive adapter based on gated units and residual blocks was introduced to achieve multi-scale feature alignment and redundancy suppression, and a multimodal global feature dynamic fusion module with self-attention was designed to adaptively adjust the weight allocation of different feature streams according to sample characteristics, thereby obtaining more consistent and complementary fused representations. Specifically, point clouds were rendered into multi-view depth maps, and features were extracted using both the original CLIP visual encoder and a CLIP encoder pretrained on depth maps, combined with point-cloud geometric features. After processing through the adaptive adapter, these features were fed into the attention-based fusion module and aligned with semantic features extracted by the CLIP text encoder for classification. In addition, contrastive learning loss, multi-view and geometric perturbation-based data augmentation strategies, and a memory-replay mechanism were incorporated to effectively mitigate overfitting and forgetting under few-shot conditions. Experiments on ShapeNet, ModelNet, and CO3D demonstrated that the proposed method consistently achieved higher accuracy across incremental stages compared with existing 3D FSCIL approaches, while significantly reducing both relative accuracy drop rates and maximum stage fluctuations.

Key words: 3D point cloud, incremental learning, few-shot learning, 3D classification, pre-trained model

CLC Number: