Multimodal cross-decoupling for few-shot learning
CSTR:
Author:
Affiliation:

(1. School of Electrical and Information Engineering, Tianjin University, Tianjin 300072, China;2. College of Information Science & Electronic Engineering, Zhejiang University, Hangzhou 310027, China)

Clc Number:

TP18

Fund Project:

  • Article
  • |
  • Figures
  • |
  • Metrics
  • |
  • Reference
  • |
  • Related
  • |
  • Cited by
  • |
  • Materials
  • |
  • Comments
    Abstract:

    Current multi-modal few-shot learning methods overlook the impact of inter-attribute differences on accurately recognizing sample categories. To address this problem, a multimodal cross-decoupling method was proposed which could decouple semantic features with different attributes and reconstruct the essential category features of samples, aiming to alleviate the impact of category attribute differences on category discrimination. Extensive experiments on two benchmark few-shot datasets MIT-States and C-GQA with large attribute discrepancy indicates that the proposed method outperforms the existing approaches, which fully verifies its effectiveness, indicating that the multimodal cross-decoupling few-shot learning method can improve the classification performance of identifying few test samples.

    Reference
    Related
    Cited by
Get Citation

JI Zhong, WANG Sidi, YU Yunlong. Multimodal cross-decoupling for few-shot learning[J]. Journal of National University of Defense Technology,2024,46(1):12-21.

Copy
Share
Article Metrics
  • Abstract:
  • PDF:
  • HTML:
  • Cited by:
History
  • Received:June 21,2022
  • Revised:
  • Adopted:
  • Online: January 28,2024
  • Published: February 28,2024
Article QR Code