JOURNAL ARTICLE

Embrace Smaller Attention: Efficient Cross-Modal Matching with Dual Gated Attention Fusion

Abstract

Cross-modal matching is one of the most fundamental and widely studied tasks in the field of data science. To have a better understanding of the complicated cross-modal correspondences, the powerful attention mechanism has been widely used recently. In this paper, we propose a novel Dual Gated Attention Fusion (DGAF) unit to save cross-modal matching from heavy attention computation. Specifically, the attention unit in the main information flow is alternated to a single-head low-dimension light-weighted attention bypass which serves as a gate to selectively cast away noise in both modality. To strengthen the interaction between modalities, an auxiliary memory unit is appended. A gated memory fusion unit is designed to fuse the memorized inter-modality information into both modality streams. Extensive experiments on two benchmark datasets show that the proposed DGAF achieves good balance between the efficiency and the effectiveness.

Keywords:
Computer science Modality (human–computer interaction) Benchmark (surveying) Matching (statistics) Fuse (electrical) Artificial intelligence Sensor fusion Modal Dual (grammatical number) Dimension (graph theory) Computation Modalities Algorithm Engineering Mathematics

Metrics

1
Cited By
0.18
FWCI (Field Weighted Citation Impact)
34
Refs
0.38
Citation Normalized Percentile
Is in top 1%
Is in top 10%

Citation History

Topics

Multimodal Machine Learning Applications
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Advanced Image and Video Retrieval Techniques
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
Visual Attention and Saliency Detection
Physical Sciences →  Computer Science →  Computer Vision and Pattern Recognition
© 2026 ScienceGate Book Chapters — All rights reserved.